OpenAi has been busy. First, they have some tutorials about interfacing with document collections using embeddings. Looks like a simpler version of GPT-Index
Second, they wrote up a report on using LLMs for misinformation and what to do about that:
Generative Language Models and Automated Influence Operations: Emerging Threats and Potential Mitigations
- Generative language models have improved drastically, and can now produce realistic text outputs that are difficult to distinguish from human-written content. For malicious actors, these language models bring the promise of automating the creation of convincing and misleading text for use in influence operations. This report assesses how language models might change influence operations in the future, and what steps can be taken to mitigate this threat. We lay out possible changes to the actors, behaviors, and content of online influence operations, and provide a framework for stages of the language model-to-influence operations pipeline that mitigations could target (model construction, model access, content dissemination, and belief formation). While no reasonable mitigation can be expected to fully prevent the threat of AI-enabled influence operations, a combination of multiple mitigations may make an important difference.
Journalistic Lessons for the Algorithmic Age
- At The Markup we pioneered an array of scientifically inspired methods that used automation and computational power to supercharge our journalism. Reflecting on our work, I came up with 10 of the most important lessons I’ve learned using this approach.
- Proofing chapters. Finished up to chapter 10. Minor tweaks