LeCun, Y. Testimony before the U.S. Senate Select Committee on Intelligence (2023).
Luitse, D. & Denkena, W. The great transformer: examining the role of large language models in the political economy of AI. Big Data Soc. 8, 20539517211047734 (2021).
Rikap, C. Same End By Different Means: Google, Amazon, Microsoft and Meta’s Strategies to Organize Their Frontier AI Innovation Systems. City, University of London https://www.cdh.cam.ac.uk/wp-content/uploads/2023/06/8.-Rikap-2023-Same-end-different-means-longer-version-CITYPERC.pdf (2023).
Whittaker, M. The steep cost of capture. Interactions 28, 50–55 (2021). Through a historical analysis of the military’s influence over AI of the past, this piece shows how AI of the present depends on resources held by a few companies, thus allowing them to capture AI research agendas.
Luitse, D. Platform power in AI: the evolution of cloud infrastructures in the political economy of artificial intelligence. Internet Policy Rev. 13, 1–44 (2024).
O’Brien, M. White House wades into debate on ‘open’ versus ‘closed’ artificial intelligence systems. Associated Press https://apnews.com/article/ai-executive-order-biden-opensource-models-1c42092e55729d731d246440094f7fed (2024).
Staff in the Office of Technology. On Open-Weights Foundation Models. Federal Trade Commission https://www.ftc.gov/policy/advocacy-research/tech-at-ftc/2024/07/open-weights-foundation-models (2024).
Competition & Markets Authority. CMA AI strategic update. UK Government https://www.gov.uk/government/publications/cma-ai-strategic-update/cma-ai-strategic-update (2024).
Bommasani, R. et al. Considerations for Governing Open Foundation Models. Stanford University https://hai.stanford.edu/issue-brief-considerations-governing-open-foundation-models (2023).
Karabus, J. GitHub CEO says EU AI Act shouldn’t apply to open source devs. The Register https://www.theregister.com/2023/02/07/github_ceo_ai_act/ (2023).
Deutsch, J. OpenAI Backs Idea of Requiring Licenses for Advanced AI Systems. Bloomberg https://www.bloomberg.com/news/articles/2023-07-20/internal-policy-memo-shows-how-openai-is-willing-to-be-regulated (2023).
Kapoor, S. et al. Position: On the societal impact of open foundation models. Proc. Machine Learning Res. 235, 23082–23104 (2024).
Seger, E. et al. Open-sourcing highly capable foundation models: an evaluation of risks, benefits, and alternative methods for pursuing open-source objectives. Preprint at http://arxiv.org/abs/2311.09227 (2023).
Bommasani, R. et al. The Foundation Model Transparency Index. Preprint at https://doi.org/10.48550/arXiv.2310.12941 (2023).
Liesenfeld, A., Lopez, A. & Dingemanse, M. in Proc. 5th International Conference on Conversational User Interfaces 1–6 https://doi.org/10.1145/3571884.3604316 (ACM, 2023).
Suchman, L. The uncontroversial ‘thingness’ of AI. Big Data Soc. 10, 20539517231206794 (2023). An important short piece problematizing the stability of the notion of ‘AI’.
Gent, E. The tech industry can’t agree on what open-source AI means. That’s a problem. MIT Technology Review https://www.technologyreview.com/2024/03/25/1090111/tech-industry-open-source-ai-definition-problem/ (2024).
Open Source Initiative. The Open Source AI Definition – 1.0. Open Source Initiative https://opensource.org/ai/open-source-ai-definition (2024).
Coleman, E. G. Coding Freedom: The Ethics and Aesthetics of Hacking (Princeton Univ. Press, 2013). An ethnography examining the politics and ethics of free and open source software communities.
Kelty, C. M. Two Bits: The Cultural Significance of Free Software (Duke Univ. Press, 2008).
Meneely, A. & Williams, L. in Proc. 16th ACM Conference on Computer and Communications Security 453–462 https://doi.org/10.1145/1653662.1653717 (ACM, 2009).
Rosenberg, J. The meaning of open. Google Public Policy Blog https://publicpolicy.googleblog.com/2009/12/meaning-of-open.html (2009).
Alexy, O. Free Revealing: How Firms Can Profit From Being Open (Springer, 2009).
IBM. A strong history and commitment to open source. IBM https://www.ibm.com/opensource/story/.
AT.40099 – Google Android. European Commission https://competition-cases.ec.europa.eu/cases/AT.40099 (2015).
Barr, J. Amazon DocumentDB (with MongoDB Compatibility): Fast, Scalable, and Highly Available. AWS News Blog https://aws.amazon.com/blogs/aws/new-amazon-documentdb-with-mongodb-compatibility-fast-scalable-and-highly-available/ (2019).
Staff in the Office of Technology. Cloud Computing: Taking Stock and Looking Ahead. Federal Trade Commission https://www.ftc.gov/news-events/events/2023/05/cloud-computing-taking-stock-looking-ahead (2023).
McLaughlin, K. & Gardizy, A. After Years of Resistance, AWS Opens Checkbook for Open-Source Providers. The Information https://www.theinformation.com/articles/after-years-of-resistance-aws-opens-checkbook-for-open-source-providers (2023).
Asay, M. & Mehla, A. MongoDB and AWS Expand Global Collaboration. MongoDB https://www.mongodb.com/blog/post/mongodb-aws-expand-global-collaboration (2022).
Zuckerberg, M. Open Source AI Is the Path Forward. Meta https://about.fb.com/news/2024/07/open-source-ai-is-the-path-forward/ (2024).
Zuckerberg, M. & Dorell, K. Fourth Quarter 2023 Results conference call (2024).
Ojewale, V., Steed, R., Vecchione, B., Birhane, A. & Raji, I. D. Towards AI accountability infrastructure: gaps and opportunities in AI audit tooling. Preprint at https://doi.org/10.48550/arXiv.2402.17861 (2024).
Gururaja, S., Bertsch, A., Na, C., Widder, D. & Strubell, E. in Proc. 2023 Conference on Empirical Methods in Natural Language Processing (eds Bouamor, H., Pino, J. & Bali, K.) 13310–13325 https://doi.org/10.18653/v1/2023.emnlp-main.822 (Association for Computational Linguistics, 2023). This is an interview study of academic NLP researchers on the incentives governing their field and on the effect of corporate power in their practice.
Khan, L. M. Sources of tech platform power. Georgetown Law Technol. Rev. 2, 325 (2018).
Vipra, J. & Korinek, A. Market concentration implications of foundation models: the invisible hand of ChatGPT. Brookings Center on Regulation and Markets https://www.brookings.edu/articles/market-concentration-implications-of-foundation-models-the-invisible-hand-of-chatgpt/ (2023).
Ezrachi, A. & Stucke, M. E. Virtual Competition: The Promise and Perils of the Algorithm-Driven Economy (Harvard Univ. Press, 2019).
OECD. Artificial intelligence, data and competition. OECD https://www.oecd.org/en/publications/artificial-intelligence-data-and-competition_e7e88884-en.html (2024).
Competition and Markets Authority, UK Government. AI Foundation Models Technical Update Report. UK Government https://www.gov.uk/cma-cases/ai-foundation-models-initial-review (2024).
West, S. M. & Kak, A. AI Now 2023 Landscape: Confronting Tech Power. AI Now Institute https://ainowinstitute.org/2023-landscape (2023).
Nathan, A., Grimberg, J. & Rhodes, A. Gen AI: Too Much Spend, Too Little Benefit? Goldman Sachs https://www.goldmansachs.com/intelligence/pages/gs-research/gen-ai-too-much-spend-too-little-benefit/report.pdf (2024).
Solaiman, I. in Proc. 2023 ACM Conference on Fairness, Accountability, and Transparency 111–122 https://doi.org/10.1145/3593013.3593981 (ACM, 2023). This is a key piece that was the first to argue that AI is not ‘binary’ in an AI context.
Meta. Introducing Meta Llama 3: The most capable openly available LLM to date. Meta AI https://ai.meta.com/blog/meta-llama-3/ (2024).
Meta. Meta Llama 3 License. Meta https://llama.meta.com/llama3/license/ (2024).
Maffulli, S. Meta’s LLaMa 2 license is not Open Source. Open Source Initiative https://opensource.org/blog/metas-llama-2-license-is-not-open-source (2023).
Nolan, M. Llama and ChatGPT Are Not Open-Source. IEEE Spectrum https://spectrum.ieee.org/open-source-llm-not-open (2023).
Hull, C. Is Llama 2 open source? No – and perhaps we need a new definition of open… OpenSource Connections https://opensourceconnections.com/blog/2023/07/19/is-llama-2-open-source-no-and-perhaps-we-need-a-new-definition-of-open/ (2023).
Buchanan, B. The AI Triad and What It Means for National Security Strategy. Center for Security and Emerging Technology https://cset.georgetown.edu/publication/the-ai-triad-and-what-it-means-for-national-security-strategy/ (2020).
Musser, M., Gelles, R., Kinoshita, R., Aiken, C. & Lohn, A. The Main Resource Is the Human. Center for Security and Emerging Technology https://cset.georgetown.edu/publication/the-main-resource-is-the-human/ (2023).
Vipra, J. & West, S. M. Computational Power and AI. AI Now Institute https://ainowinstitute.org/publication/policy/compute-and-ai (2023).
Dettmers, T., Pagnoni, A., Holtzman, A. & Zettlemoyer, L. QLoRA: Efficient Finetuning of Quantized LLMs. Proc. 37th Int. Conf. Neural Inf. Process. (NIPS ’23) Article 441, 10088–10115 (Curran Assoc., Red Hook, NY, 2024).
Birhane, A., Prabhu, V., Han, S., Boddeti, V. N. & Luccioni, A. S. Into the LAION’s Den: Investigating Hate in Multimodal Datasets. 37th Conf. Neural Inf. Process. Syst. Datasets Benchmarks Track https://openreview.net/forum?id=6URyQ9QhYv (2023).
Gerchick, M. et al. in Proc. 2023 ACM Conference on Fairness, Accountability, and Transparency 1292–1310 https://doi.org/10.1145/3593013.3594081 (ACM, 2023).
Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 1, 206–215 (2019).
Lipton, Z. C. The mythos of model interpretability: in machine learning, the concept of interpretability is both important and slippery. Queue 16, 31–57 (2018).
Poursabzi-Sangdeh, F., Goldstein, D. G., Hofman, J. M., Wortman Vaughan, J. W. & Wallach, H. in Proc. 2021 CHI Conference on Human Factors in Computing Systems 1–52 https://doi.org/10.1145/3411764.3445315 (ACM, 2021).
EleutherAI. License – EleutherAI/gpt-neox. GitHub https://github.com/EleutherAI/gpt-neox?tab=Apache-2.0-1-ov-file (2024).
Metz, C. Mistral, French A.I. Start-Up, Is Valued at $2 Billion in Funding Round. The New York Times https://www.nytimes.com/2023/12/10/technology/mistral-ai-funding.html (2023).
Mistral AI. https://mistral.ai/.
Boyd, E. Introducing Mistral-Large on Azure in partnership with Mistral AI. Microsoft Azure Blog https://azure.microsoft.com/en-us/blog/microsoft-and-mistral-ai-announce-new-partnership-to-accelerate-ai-innovation-and-introduce-mistral-large-first-on-azure/ (2024).
Touvron, H. et al. Llama 2: open foundation and fine-tuned chat models. Preprint at https://doi.org/10.48550/arXiv.2307.09288 (2023).
Bender, E. M., Gebru, T., McMillan-Major, A. & Shmitchell, S. in Proc. 2021 ACM Conference on Fairness, Accountability, and Transparency 610–623 https://doi.org/10.1145/3442188.3445922 (ACM, 2021). This article documents that the creation of massive training datasets renders them difficult to scrutinize and increasingly opaque.
Bertsch, A. et al. in Proc. 4th Workshop on Gender Bias in Natural Language Processing (GeBNLP) (eds Hardmeier, C., Basta, C., Costa-jussà, M. R., Stanovsky, G. & Gonen, H.) 235–243 https://doi.org/10.18653/v1/2022.gebnlp-1.24 (Association for Computational Linguistics, 2022).
Longpre, S. et al. A Pretrainer’s Guide to Training Data: Measuring the Effects of Data Age, Domain Coverage, Quality, & Toxicity. Proc. 24 Conf. N. Am. Ch. Assoc. Comp. Linguistics Hum. Language Tech. Vol. 1, 3245–3276 (Assoc. Comp. Linguistics, Mexico City, Mexico, 2023).
Widder, D. G. & Nafus, D. Dislocated accountabilities in the “AI supply chain”: modularity and developers’ notions of responsibility. Big Data Soc. 10, 20539517231177620 (2023).
Almazrouei, E. et al. The Falcon series of open language models. Preprint at https://doi.org/10.48550/arXiv.2311.16867 (2023).
The MosaicML NLP Team. Introducing MPT-7B: A New Standard for Open-Source, Commercially Usable LLMs. Databricks https://www.databricks.com/blog/mpt-7b (2023).
Scao, T. L. et al. BLOOM: a 176B-parameter open-access multilingual language model. Preprint at https://arxiv.org/abs/2211.05100 (2023).
Osman, L. & Sewell, D. Hugging Face. Contrary Research https://research.contrary.com/company/hugging-face (2022).
EleutherAI. EleutherAI – About. EleutherAI https://www.eleuther.ai/about.
Edwards, B. You can now run a GPT-3-level AI model on your laptop, phone, and Raspberry Pi. Ars Technica https://arstechnica.com/information-technology/2023/03/you-can-now-run-a-gpt-3-level-ai-model-on-your-laptop-phone-and-raspberry-pi/ (2023).
Germain, T. Stanford Researchers Take Down Alpaca AI Due to ‘Hallucinations’ and Rising Costs. Gizmodo Australia https://gizmodo.com/stanford-ai-alpaca-llama-facebook-taken-down-chatgpt-1850247570 (2023).
OpenAI. GPT-4 technical report. Preprint at http://arxiv.org/abs/2303.08774 (2023).
Anthropic. The Claude 3 Model Family: Opus, Sonnet, Haiku. Anthropic https://www-cdn.anthropic.com/de8ba9b01c9ab7cbabf5c33b80b7bbc618857627/Model_Card_Claude_3.pdf.
Mensch, A. mistralai/Mistral-7B-v0.1. Training data? Hugging Face https://huggingface.co/mistralai/Mistral-7B-v0.1/discussions/8 (2023).
Xu, H. et al. Demystifying CLIP data. Preprint at https://arxiv.org/abs/2309.16671 (2023).
Kaplan, J. et al. Scaling laws for neural language models. Preprint at https://doi.org/10.48550/arXiv.2001.08361 (2020).
Gebru, T. et al. Datasheets for datasets. Commun. ACM 64, 86–92 (2021).
GhanaNLP. About Us. GhanaNLP https://ghananlp.org/about.
Deck, A. The AI startup outperforming Google Translate in Ethiopian languages. Rest of World https://restofworld.org/2023/3-minutes-with-asmelash-teka-hadgu/ (2023).
Knibbs, K. Here’s Proof You Can Train an AI Model Without Slurping Copyrighted Content. WIRED https://www.wired.com/story/proof-you-can-train-ai-without-slurping-copyrighted-content/ (2024).
Khan, M. & Hanna, A. The subjects and stages of AI dataset development: a framework for dataset accountability. Ohio State Law J. 19, 171 (2023).
Smith, B. & Nowbar, H. Microsoft announces new Copilot Copyright Commitment for customers. Microsoft On the Issues https://blogs.microsoft.com/on-the-issues/2023/09/07/copilot-copyright-commitment-ai-legal-concerns/ (2023).
Jiang, H. H. et al. in Proc. 2023 AAAI/ACM Conference on AI, Ethics, and Society 363–374 https://doi.org/10.1145/3600211.3604681 (ACM, 2023).
Setty, R. Sarah Silverman, Authors Hit OpenAI, Meta With Copyright Suits. Bloomberg Law https://news.bloomberglaw.com/ip-law/sarah-silverman-authors-hit-openai-meta-with-copyright-suits (2023).
Browne, G. AI Is Steeped in Big Tech’s ‘Digital Colonialism’. WIRED https://www.wired.com/story/abeba-birhane-ai-datasets/ (2023).
Azunre, P. [@pazunre]. If African AI/ML researchers are not careful, this new “Open Source” movement championed by the richest global tech companies will become a mechanism for continued exploitation of our human capital and continent… [1/3]. Twitter https://twitter.com/pazunre/status/1569743778524680192 (2022).
Abebe, R. et al. in Proc. 2021 ACM Conference on Fairness, Accountability, and Transparency 329–341 https://doi.org/10.1145/3442188.3445897 (ACM, 2021).
Común, T. Resisting Data Colonialism – A Practical Intervention (Institute of Network Cultures, 2023).
Te Hiku Media. Kaitiakitanga-License/LICENSE.md. GitHub https://github.com/TeHikuMedia/Kaitiakitanga-License/blob/tumu/LICENSE.md (2018).
Vincent, J. The lawsuit that could rewrite the rules of AI copyright. The Verge https://www.theverge.com/2022/11/8/23446821/microsoft-openai-github-copilot-class-action-lawsuit-ai-copyright-violation-training-data (2022).
Gao, L. et al. The Pile: an 800GB dataset of diverse text for language modeling. Preprint at https://doi.org/10.48550/arXiv.2101.00027 (2020).
Baack, S. A in Proc. 2024 ACM Conference on Fairness, Accountability, and Transparency 2199–2208 https://doi.org/10.1145/3630106.3659033 (ACM, 2024).
Luccioni, A. S. & Viviano, J. D. What’s In The Box? A Preliminary Analysis of Undesirable Content in the Common Crawl Corpus. Proc. 59th Annu. Meeting Assoc. Comp. Linguistics and 11th Int. Joint Conf. Natural Lang. Processing Vol. 2 182–189 (Assoc. Comp. Linguistics, 2021).
Taylor, A. The Automation Charade. Logic(s) Magazine https://logicmag.io/failure/the-automation-charade/ (2018).
Williams, A., Miceli, M. & Gebru, T. The Exploited Labor Behind Artificial Intelligence. Noema Magazine https://www.noemamag.com/the-exploited-labor-behind-artificial-intelligence/ (2022).
Thylstrup, N. & Talat, Z. Detecting ‘Dirt’ and ‘Toxicity’: Rethinking Content Moderation as Pollution Behaviour. SSRN https://doi.org/10.2139/ssrn.3709719 (2020).
Christiano, P. et al. Deep Reinforcement Learning From Human Preferences. Proc. 31 Int. Conf. Neural Inf. Processing Syst. (NIPS ’17) 4302–4310 (Curran Assoc., Red Hook, NY, 2017).
Perrigo, B. Exclusive: OpenAI Used Kenyan Workers on Less Than $2 Per Hour to Make ChatGPT Less Toxic. Time.com https://time.com/6247678/openai-chatgpt-kenya-workers/ (2023).
Hao, K. The Hidden Workforce That Helped Filter Violence and Abuse Out of ChatGPT. The Wall Street Journal https://www.wsj.com/podcasts/the-journal/the-hidden-workforce-that-helped-filter-violence-and-abuse-out-of-chatgpt/ffc2427f-bdd8-47b7-9a4b-27e7267cf413 (2023).
Hao, K. & Hernández, A. P. How the AI industry profits from catastrophe. MIT Technology Review https://www.technologyreview.com/2022/04/20/1050392/ai-industry-appen-scale-data-labels/ (2022).
Perrigo, B. 150 African Workers for ChatGPT, TikTok and Facebook Vote to Unionize at Landmark Nairobi Meeting. Time.com https://time.com/6275995/chatgpt-facebook-african-workers-union/ (2023).
Mutemi, M. [@MercyMutemi]. On behalf of the young Kenyans whose lives have been ruined because they did the dirty work training the #ChatGPT algorithm, we have filed a petition to @NAssemblyKE to investigate how @OpenAI and @samasource got away with such exploitation and to urgently regulate tech work. https://t.co/9seeyGKqFM. Twitter https://twitter.com/MercyMutemi/status/1678984336996028416 (2023).
Irani, L. C. & Silberman, M. S. in Proc. SIGCHI Conference on Human Factors in Computing Systems 611–620 https://doi.org/10.1145/2470654.2470742 (ACM, 2013).
Foster, K. PyTorch vs TensorFlow: Who has More Pre-trained Deep Learning Models? HackerNoon https://hackernoon.com/pytorch-vs-tensorflow-who-has-more-pre-trained-deep-learning-models (2022).
He, H. The State of Machine Learning Frameworks in 2019. The Gradient https://thegradient.pub/state-of-ml-frameworks-2019-pytorch-dominates-research-tensorflow-dominates-industry/ (2019).
Paszke, A. et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proc. 33 Int. Conf. Neural Inf. Processing Syst. Article 721, 8026–8037 (Curran Assoc., Red Hook, NY, 2019).
Meta. Announcing the PyTorch Foundation: A new era for the cutting-edge AI framework. Meta https://ai.meta.com/blog/pytorch-foundation/ (2022).
Tarantola, A. Meta is spinning off the Pytorch framework into its own AI research foundation. Engadget https://www.engadget.com/meta-is-spinning-off-its-pytorch-ai-framework-into-its-own-research-foundation-140051987.html (2022).
PyTorch Contributors. PyTorch Governance: Maintainers. PyTorch 2.2 documentation. PyTorch https://pytorch.org/docs/stable/community/persons_of_interest.html (2023).
Metz, C. Google Just Open Sourced TensorFlow, Its Artificial Intelligence Engine. WIRED https://www.wired.com/2015/11/google-open-sources-its-artificial-intelligence-engine/ (2015).
Contributors to tensorflow/tensorflow. GitHub https://github.com/tensorflow/tensorflow/graphs/contributors (2024).
Engler, A. How open-source software shapes AI policy. Brookings https://www.brookings.edu/articles/how-open-source-software-shapes-ai-policy/ (2021).
Zuckerberg, M. & Crawford, D. First Quarter 2023 Results conference call (2023).
Cath-Speth, C. J. N. Changing Minds and Machines: A Case Study of Human Rights Advocacy in the Internet Engineering Task Force (IETF). Thesis, Univ. Oxford (2021).
Langenkamp, M. & Yue, D. N. in Proc. 2022 AAAI/ACM Conference on AI, Ethics, and Society 385–395 https://doi.org/10.1145/3514094.3534167 (ACM, 2022). This study examines the economic value of open-source AI development.
Schwartz, R., Dodge, J. & Smith, N. A. Green AI. Commun. ACM 63, 54–63 (2020).
Villalobos, P. et al. Will we run out of data? Limits of LLM scaling based on human-generated data. Preprint at http://arxiv.org/abs/2211.04325 (2024).
Luccioni, S., Jernite, Y. & Strubell, E. in Proc. 2024 ACM Conference on Fairness, Accountability, and Transparency 85–99 https://doi.org/10.1145/3630106.3658542 (ACM, 2024). This paper measures the carbon cost (and therefore, indirectly, the US$ cost) of different AI models, thus giving a scale of the resources they require to run.
Leswing, K. Nvidia dominates the AI chip market, but there’s more competition than ever. CNBC https://www.cnbc.com/2024/06/02/nvidia-dominates-the-ai-chip-market-but-theres-rising-competition-.html (2024).
Cherney, M. A. Exclusive: Behind the plot to break Nvidia’s grip on AI by targeting software. Reuters https://www.reuters.com/technology/behind-plot-break-nvidias-grip-ai-by-targeting-software-2024-03-25/ (2024).
Why do Nvidia’s chips dominate the AI market? The Economist https://www.economist.com/the-economist-explains/2024/02/27/why-do-nvidias-chips-dominate-the-ai-market (2024).
Dow, R. Shipments of graphics add-in boards decline in Q1 of 24 as the market experiences a return to seasonality. Jon Peddie Research https://www.jonpeddie.com/news/shipments-of-graphics-add-in-boards-decline-in-q1-of-24-as-the-market-experiences-a-return-to-seasonality/ (2024).
Patel, D. & Nishball, D. Google Gemini Eats The World – Gemini Smashes GPT-4 By 5X, The GPU-Poors. Semianalysis https://www.semianalysis.com/p/google-gemini-eats-the-world-gemini (2023).
Dotan, T. Microsoft Earnings Growth Accelerates on Stronger-Than-Expected Cloud Demand. The Wall Street Journal https://www.wsj.com/tech/microsoft-msft-q1-earnings-report-2024-b19e51eb (2023).
Vyas, I. AI and Machine Learning Integration in SaaS Applications. IEEE Computer Society https://www.computer.org/publications/tech-news/trends/ai-and-machine-learning-integration/ (2023).
Lowe, L. AI, Startups, & Competition: Shaping California’s Tech Future. Chaindesk https://www.chaindesk.ai/tools/youtube-summarizer/ai-startups-and-competition-shaping-california-s-tech-future-pieVtTrbDBs (2024).