We conjecture that hidden state vectors corresponding to individual input tokens encode information sufficient to accurately predict several tokens ahead. More concretely, in this paper we ask: Given a hidden (internal) representation of a single token at position t in an input, can we reliably anticipate the tokens that will appear at positions ≥t+2? To test this, we measure linear approximation and causal intervention methods in GPT-J-6B to evaluate the degree to which individual hidden states in the network contain signal rich enough to predict future hidden states and, ultimately, token outputs. We find that, at some layers, we can approximate a model’s output with more than 48% accuracy with respect to its prediction of subsequent tokens through a single hidden state. Finally we present a “Future Lens” visualization that uses these methods to create a new view of transformer states.
This looks like a good paper to extend the NNM work. They are looking at activations in different layers and using them to work out a trajectory. Need to dig into further.
Happy 1st birthday to ChatGPT, for those who celebrate
Sent a note to Scott Shapiro
Sent emails to the people who registered for ContextTest but didn’t start the study
Alibaba opensources our Qwenseries, now including Qwen, the base language models, namely Qwen-1.8B, Qwen-7B, Qwen-14B, and Qwen-72B, as well as Qwen-Chat, the chat models, namely Qwen-1.8B-Chat, Qwen-7B-Chat, Qwen-14B-Chat, and Qwen-72B-Chat. Links are on the above table. Click them and check the model cards. Also, we release the technical report. Please click the paper link and check it out!
Qwen-Agentis a framework for harnessing the tool usage, planning, and memory capabilities of the open-source language model Qwen. Building upon Qwen-Agent, we have developed a Chrome browser extension called BrowserQwen, which has key features such as:
You can discuss with Qwen regarding the current webpage or PDF document.
It records the web pages and PDF materials that you have browsed, with your permission. It helps you quickly understand the contents of multiple pages, summarize your browsing content, and eliminate tedious writing tasks.
It supports plugin integration, including Code Interpreter for math problem solving and data visualization.
It supports uploading PDF, Word, PPT and other types of documents for multi document Q&A.
Freaked myself out when I saw one of my variations on the Cozy Bear Podesta email. That’s such an effective technique!
SBIRs
Q7 report – Finished Task 2
Check with Aaron, but aim to submit the slide deck by COB – getting approval
Put the GPT content in the Killer Apps project for later reference. – done
Wound up having to do more self-evaluation
Had a nice chat with Aaron about prompt swarms, or maybe synthetic organisms (sorgs?). We’re going to look into setting up agents to run the Grog dungeon. I’m really curious about how they will handle the Trolley problem
GPT Agents
Send email to Scott Shapiro
Alden Meeting – went well. Some interesting stuff about how models become more negative as the temp goes up. I’m thinking that there could be something like a phase change?
Had a good (in person!) meeting with Tyler. I’m not sure that he buys the chemistry metaphor, but he does agree that generative models work in a bounded domain. One thought is to use an LLM to do chemistry. Train it on SMILES notation, and let it do chemical reactions. It could be a smaller, GPT-2 sized model too. For that matter, it would be straightforward to write a SMILES expander like I did with the GPT-2 chess experiments, so that the model only needs to be finetuned. Other aspects, like energy would have to be included to see if the model could produce new “trajectories” through the reaction space. The other option is to use the chess model again, since that’s another bounded domain that the model can clearly work with.
Built a mockup of the infamous Podesta phishing email
Here’s the html:
<!DOCTYPE html>
<html>
<head>
<title>Google Security Alert</title>
<style>
body {
font-family: 'Product Sans', sans-serif;
}
</style>
</head>
<body>
<div style="width: 100%; max-width: 600px; margin: 0
auto; padding: 20px">
<img style="width: 30%;" src="https://www.google.com/images/branding/googlelogo/1x/googlelogo_color_272x92dp.png" alt="Google Security Alert">
<h2 style="background-color: red; color: white; padding: 20px">Someone has your password</h2>
<p>Hi John</p>
<p>Someone just used your password to try to sign into your Google Account <a href="mailto:john.podesta@gmail.com">john.podesta@gmail.com</a>.
<h3>Details</h3>
<ul>
<li>Saturday, 19 March, 8:34:30 UTC</li>
<li>IP Address 134.249.139.239</li>
<li>Location: Ukraine</li>
</ul>
<p>Google stopped this sign-in attempt. You should change your password immediately</p>
<div style="text-align: left;">
<a href="https://bit.ly/1PibSUO" style="background-color:#4c8bf5; color: white; padding: 10px; text-decoration: none; font-weight: bold;">
Change Password</a>
</div>
</div>
</body>
</html>
The GPT-4, with the system prompt “You are a cybersecurity expert with a deep understanding of the human behaviors that hackers use to exploit their targets. You explain these techniques clearly and concisely,” has no problem recognizing it with zero shots of training:
Caution: The email you received, claiming to be a security alert from Google, raises several red flags and should be treated with suspicion. The use of a shortened URL for password reset, the inclusion of an email link, and an alarming visual design are techniques typically employed by phishers to create urgency and prompt hasty action. The absence of personalized information and the use of a non-Google domain for the logo image are also concerning signs. Do not click any links within this email. Instead, for your security, access your Google account directly through a new browser tab and verify any security issues from there. Remember to be vigilant and report this email as a potential phishing attempt to help protect others.
Finished the slide deck. I added a section about prompt swarms learning for defense against social AI attacks. Need to set up an Overleaf for that project.
Transferred the slide deck to the ASRC format. Need to send it to MARCOM for blessings
Working on the ETF presentation because I think it won’t be ready otherwise, and it needs a lot of work.
Changed the title to “Killer Apps,” which has a nice ring to it
Adding a section about Spear Phishing using the Podesta email
Working on a mockup of how an AI could spot things like this. Ideally, set up a model that can do this. The rules and examples should be straigtforward.
Tie into the collective learning prompt swarm concept for adaptation.
Magnesium seems to be helping my power and maybe a bit with cramping? Might have to adjust my diet a bit? Ask for a bloodwork check in my physical for sure
SBIRs
Got a chance to work on the ETF slides a lot. I’m beginning to think that the OODA loop section may be a blind ally, since we’re talking about societal scales. Maybe bring it in as a way to look at the problem, rather than defending it?
Spent the entire afternoon on IRAD schedules and budgeting
Had a thought about using the new GPT agents. I think they can be best used one chapter at a time when writing. First, all the assets at once exceed the 20-item limit. Second, the model can’t do large-scale contextualization.
Conflict has become increasingly prevalent in developing countries, and the role of social media platforms in exacerbating these conflicts cannot be ignored. Peacebuilders who focus on promoting inclusive and sustainable peace in war-torn countries are confronted with numerous challenges. The effects of disinformation, misinformation, polarization, online harassment, and the use of digital media as political weapons have been extensively examined in the context of the USA and Europe. The situation in developing nations such as Ethiopia, characterized by a significant digital divide, ethnic polarization, ethnification of mass media, and limited access to digital media for most of the population, remains understudied. Within such conflicts, women play a crucial role as they are often the last barrier for economic collapse, but likewise are specifically targeted in conflicts. Women are historically marginalized all over the world, but especially in the context of developing nations, Ethiopia in particular. The two-year-long Ethiopian civil war showed that women were more seriously affected by the war than their male counterparts. From the different national and international media reports, we have learned that mass displacement, sexual harassment, using rape as a tool of war, group rape, etc. challenged the lives of women in Ethiopia. Women paid the price of the war more than men in the country. Thus, incorporating women’s voices, perspectives, and experiences is paramount for inclusive and sustainable peacebuilding. Our research proposal seeks to explore the impact of social media on offline unrest, specifically its effects on women, and provide viable solutions to peacebuilders. We focus on building a pipeline for digital peacebuilding, including the potential use of AI tools such as large language models (LLMs) like Google’s PaLM, as automated classifiers. Data will be collected from popular social media platforms in Ethiopia, with a focus on addressing the issue of polarization affecting women. The project will further apply different NLP techniques such as topic clustering, named entity recognition, sentiment analysis, and hate speech detection with machine learning approaches. The development of such a pipeline facilitates the works of peacebuilders and aims to reduce the marginalization of women’s voices and perspectives in the peace-building process. This could lead to develop a toolchain that can be applied in a similar war-torn country such as Yemen, Libya, Sudan, etc.
GPT Agents
Tyler wrote back. Need to schedule something for next week
SBIRs
Working on the ETF deck. It’s expanding a bit too much maybe, but I can edit later
Need to fold Zac’s paragraphs into the notes.
The data is not up yet on the server. Rukan says probably not until Wednesday COB
9:00 Standup
2:30 AI Ethics? Nope
3:30 USNA – they are all over the place. Asked them to clarify their research questions and methods to answer them. We’re going to have more formalized presentations than an ad-hoc problem-solving session
Large language models (LMs) are currently trained to predict tokens given document prefixes, enabling them to directly perform long-form generation and prompting-style tasks which can be reduced to document completion. Existing pretraining pipelines train LMs by concatenating random sets of short documents to create input contexts but the prior documents provide no signal for predicting the next document. We instead present In-Context Pretraining, a new approach where language models are pretrained on a sequence of related documents, thereby explicitly encouraging them to read and reason across document boundaries. We can do In-Context Pretraining by simply changing the document ordering so that each context contains related documents, and directly applying existing pretraining pipelines. However, this document sorting problem is challenging. There are billions of documents and we would like the sort to maximize contextual similarity for every document without repeating any data. To do this, we introduce approximate algorithms for finding related documents with efficient nearest neighbor search and constructing coherent input contexts with a graph traversal algorithm. Our experiments show In-Context Pretraining offers a simple and scalable approach to significantly enhance LMs’performance: we see notable improvements in tasks that require more complex contextual reasoning, including in-context learning (+8%), reading comprehension (+15%), faithfulness to previous contexts (+16%), long-context reasoning (+5%), and retrieval augmentation (+9%).
NVIDIA® cuOpt™ is a world-record-breaking accelerated optimization engine. cuOpt helps teams solve complex routing problems with multiple constraints and deliver new capabilities, like dynamic rerouting, job scheduling, and robotic simulations, with subsecond solver response time.
Sprint review and planning yesterday. Mostly BD, which is very frustrating
Finished(?) the white paper and sent it to Orest. Nope, gotta turn it into the right format. So it turns out that there is a way of using a MSWord template, but it doesn’t seem to work on my file. Trying to send the broken version rather than adding in everything by hand.
Fixing by hand. Starting at 10:30. finished at 6:00 or so, working around meetings
Going to make a LaTeX template next time I write a white paper.
9:00 standup
2:00 BMD Bi-weekly
3:30 – 5:00 AIMSS proposal prep? Still not sure what this is.
GPT Agents
Still not seeing any invitations go out. Ping Greg today and ask him to contact his chair – done
Got a game from Antonio to test and review – done
Pulled off all mentions of Supabase from the informed consent, which goes to show that even the IRB doesn’t read them.
Finished reviews for IUI 2024 and submitted! Days ahead of schedule!
You must be logged in to post a comment.