The Economic Impact of Language Models on Influence Operations
Examining the cost savings of using language models for manipulation and influence.
― 5 min read
Table of Contents
Recent advancements in large Language Models (LLMs) have raised concerns about their potential misuse, particularly in spreading false or misleading information. Many wonder how these tools could help groups aiming to influence public opinion or manipulate information on a large scale. This article looks at the cost aspects of using LLMs for such Influence Operations, focusing on how they could save money for those who misuse them.
Understanding Language Models
Language models are computer programs designed to generate human-like text. They can produce coherent sentences and paragraphs based on prompts or themes provided by users. As these models improve in quality and speed, they become more attractive for various applications, including the creation of deceptive or misleading content for social media campaigns.
Costs of Traditional Influence Operations
To grasp the economic benefits of using LLMs, it is essential to first look at the costs involved in conventional influence operations that rely primarily on human efforts.
Labor Costs
Human authors need to be paid for their work, and their productivity can vary significantly. Analyzing wages and output rates of individuals involved in such operations provides insight into the baseline costs of generating content. For example, individuals working in social media positions often earn modest hourly wages, which can accumulate rapidly over time, especially for large campaigns that generate thousands of posts.
Infrastructure Costs
Besides labor, organizations running influence operations must also maintain infrastructure. This includes managing accounts on social media platforms, running campaigns, and even ensuring tools are in place to analyze the effectiveness of their content. All of these components add to the total cost burden of traditional influence operations, making it clear that any way to cut these costs would be appealing.
LLMs and Cost Savings
With the backdrop of traditional costs in mind, we delve into how leveraging LLMs can lead to significant savings. The following sections explore different scenarios to illustrate potential savings from using LLMs.
Fully Automated Content Generation
In this scenario, an organization uses a language model to generate content without human intervention. The only costs incurred would be those associated with operating the model. The expenses for running these models depend on their size and performance, but they are often much lower than equivalent human labor costs.
Human-Machine Teams
In real-world applications, organizations may not opt for fully automated systems but rather combine both human and AI efforts. In this framework, a language model generates content which is then reviewed and approved by a human operator. This human-machine team can save money depending on how much faster the team can produce usable content compared to relying solely on human authorship.
The potential for savings is particularly notable if the language model produces outputs that are nearly ready for posting with minimal edits. If a language model can generate a sufficiently high percentage of usable content, the combined costs of labor and operating the model can become much more cost-effective than traditional methods.
Monitoring and Controls
As concerns about misuse grow, so too do calls for monitoring the use of LLMs. Organizations that own these models can implement controls to track how they are being used and prevent misuse. However, it’s important to understand how these controls can impact cost calculations for potential misuse.
Deterrent Effects
If a language model requires user registration and monitoring, potential users may need to factor in the costs associated with evading such controls. When users are likely to face penalties or increased operational costs due to monitoring, this can deter their engagement with the model.
Open Source vs. Private Models
Another aspect to consider is the choice between Open-source models and private models with monitoring controls. Private models may have safeguards to prevent misuse, but open-source options offer greater freedom for exploitation. If an operator can choose between a monitored private model or a free-to-download open-source model, the latter may often be preferred.
Cost Comparisons
The flexibility of open-source models means that operators can generate content without the overhead of monitoring or penalties, leading to even greater savings. This scenario raises questions about the effectiveness of monitoring programs if users can easily switch to alternative systems that have fewer restrictions.
The Role of Nation-States
Many nation-state actors have been accused of using influence campaigns extensively. For these entities, the stakes are high, and their operations often span millions of social media posts over extended periods.
Economic Justifications
Despite the massive scale, national governments may find it economically unviable to invest in creating their own language models simply for influence operations. Most often, nations rely on existing models, either through open-source options or external commercial providers.
Conclusion
The use of large language models in influence operations presents a complex economic picture. While they offer significant cost savings compared to traditional methods, the challenges posed by monitoring controls and the preference for open-source models highlight ongoing tensions in managing these technologies. As the world becomes increasingly reliant on digital communication, understanding and managing the economic implications of these systems is crucial for maintaining public trust and integrity in information dissemination.
Future Considerations
As technology continues to evolve, so too will the methods of influence operations. Research into AI-generated content, its effectiveness, and the economic implications for various stakeholders will remain crucial.
The interplay between cost, efficacy, and ethical considerations surrounding the use of language models in public discourse underscores the need for continued dialogue and understanding.
Title: A Cost Analysis of Generative Language Models and Influence Operations
Abstract: Despite speculation that recent large language models (LLMs) are likely to be used maliciously to improve the quality or scale of influence operations, uncertainty persists regarding the economic value that LLMs offer propagandists. This research constructs a model of costs facing propagandists for content generation at scale and analyzes (1) the potential savings that LLMs could offer propagandists, (2) the potential deterrent effect of monitoring controls on API-accessible LLMs, and (3) the optimal strategy for propagandists choosing between multiple private and/or open source LLMs when conducting influence operations. Primary results suggest that LLMs need only produce usable outputs with relatively low reliability (roughly 25%) to offer cost savings to propagandists, that the potential reduction in content generation costs can be quite high (up to 70% for a highly reliable model), and that monitoring capabilities have sharply limited cost imposition effects when alternative open source models are available. In addition, these results suggest that nation-states -- even those conducting many large-scale influence operations per year -- are unlikely to benefit economically from training custom LLMs specifically for use in influence operations.
Authors: Micah Musser
Last Update: 2023-08-07 00:00:00
Language: English
Source URL: https://arxiv.org/abs/2308.03740
Source PDF: https://arxiv.org/pdf/2308.03740
Licence: https://creativecommons.org/licenses/by/4.0/
Changes: This summary was created with assistance from AI and may have inaccuracies. For accurate information, please refer to the original source documents linked here.
Thank you to arxiv for use of its open access interoperability.
Reference Links
- https://github.com/georgetown-cset/disinfo-costs
- https://cset.georgetown.edu/article/how-much-money-could-large-language-models-save-propagandists/
- https://www.npr.org/2022/03/16/1087062648/deepfake-video-zelenskyy-experts-war-manipulation-ukraine-russia
- https://openreview.net/pdf?id=wysXxmukfCA
- https://crestresearch.ac.uk/resources/artificial-intelligence-and-extremism-the-threat-of-language-models/
- https://arxiv.org/abs/2112.05224
- https://democracyreporting.s3.eu-central-1.amazonaws.com/images/6331fc834bcd1.pdf
- https://demtech.oii.ox.ac.uk/research/posts/industrialized-disinformation/
- https://arxiv.org/abs/2005.14165
- https://www.nber.org/system/files/working_papers/w31161/w31161.pdf
- https://www.nber.org/system/files/working
- https://cset.georgetown.edu/publication/truth-lies-and-automation/
- https://theconversation.com/ai-scam-calls-imitating-familiar-voices-are-a-growing-problem-heres-how-they-work-208221
- https://www.mcafee.com/blogs/privacy-identity-protection/artificial-imposters-cybercriminals-turn-to-ai-voice-cloning-for-a-new-breed-of-scam/
- https://arxiv.org/abs/2107.03374
- https://cohere.ai/pricing
- https://openai.com/blog/best-practices-for-deploying-language-models/
- https://cyber.fsi.stanford.edu/io/publication/one-topic-two-networks
- https://www.nature.com/articles/s41467-022-35576-9
- https://arxiv.org/abs/2303.10130
- https://www.nytimes.com/2021/07/25/world/europe/disinformation-social-media.html
- https://www.vice.com/en/article/7kbjny/russia-cyber-front-z-telegram
- https://www.youtube.com/watch?v=ev07nlTBZ3Q
- https://cdn.openai.com/papers/forecasting-misuse.pdf
- https://foreignpolicy.com/2021/12/15/china-twitter-trolls-ccp-influence-operations-astroturfing/
- https://cyber.fsi.stanford.edu/io/news/sio-aug-22-takedowns
- https://www.youtube.com/watch?v=JJZObKWG8ok
- https://www.pcguide.com/apps/gpt-3-cost/
- https://github.blog/2022-09-07-research-quantifying-github-copilots-impact-on-developer-productivity-and-happiness/
- https://doi.org/10.1017/S0003055417000144
- https://arxiv.org/abs/2301.10226
- https://archive.ph/TB4Xw
- https://www.nber.org/system/files/working_papers/w30957/w30957.pdf
- https://doi.org/10.1017/XPS.2020.37
- https://write.as/sethlazar/genb
- https://blog.eleuther.ai/why-release-a-large-language-model/
- https://arxiv.org/abs/2207.13825
- https://www.buzzfeednews.com/article/janelytvynenko/job-ads-for-russian-troll-factory
- https://www.scientificamerican.com/article/detecting-deepfakes1/
- https://www.mandiant.com/resources/blog/prc-dragonbridge-influence-elections
- https://www.middlebury.edu/institute/sites/www.middlebury.edu.institute/files/2020-09/gpt3-article.pdf
- https://arxiv.org/abs/2302.06716
- https://www.brookings.edu/blog/order-from-chaos/2018/05/25/the-west-is-ill-prepared-for-the-wave-of-deep-fakes-that-artificial-intelligence-could-unleash/
- https://arxiv.org/abs/2305.12050
- https://aisnakeoil.substack.com/p/the-llama-is-out-of-the-bag-should
- https://archive.is/UoMM9
- https://www.newsweek.com/fact-check-photo-putin-his-knees-front-chinas-xi-1789498
- https://www.science.org/doi/10.1126/science.adh2586
- https://openai.com/blog/dall-e-now-available-without-waitlist/
- https://openai.com/blog/api-no-waitlist/
- https://openai.com/api/pricing/
- https://mlmac.io/
- https://labs.withsecure.com/publications/creatively-malicious-prompt-engineering
- https://www.mosaicml.com/blog/mosaicbert
- https://openai.com/blog/better-language-models/
- https://www.independent.co.uk/news/world/americas/us-politics/trump-deepfake-arrest-twitter-ai-b2307470.html
- https://www.nytimes.com/2023/02/07/technology/artificial-intelligence-training-deepfake.html
- https://www.buzzfeednews.com/article/maxseddon/documents-show-how-russias-troll-army-hit-america
- https://cset.georgetown.edu/publication/ai-and-the-future-of-disinformation-campaigns/
- https://cset.georgetown.edu/publication/ai-and-the-future-of-disinformation-campaigns-2/
- https://arxiv.org/abs/2201.05159
- https://www.wired.com/story/generative-ai-systems-arent-just-open-or-closed-source/
- https://arxiv.org/abs/2302.04844
- https://arxiv.org/abs/1908.09203
- https://benjaminstrick.com/twitter-analysis-identifying-a-pro-bjp-influence-operation-in-india/
- https://ai.googleblog.com/2022/07/ml-enhanced-code-completion-improves.html
- https://doi.org/10.48550/arXiv.2102.02503
- https://crfm.stanford.edu/2023/03/13/alpaca.html
- https://www.washingtonpost.com/technology/interactive/2022/artificial-intelligence-images-dall-e/
- https://transparency.twitter.com/en/reports/information-operations.html
- https://arxiv.org/abs/2112.04359