Computerspeak by Alexandru Voica

Computerspeak by Alexandru Voica

Share this post

Computerspeak by Alexandru Voica
Computerspeak by Alexandru Voica
Open models are great but they have a big catch; Synthesia releases Expressive Avatars; generative AI arrives in healthcare; AI assistants and busywork.

Open models are great but they have a big catch; Synthesia releases Expressive Avatars; generative AI arrives in healthcare; AI assistants and busywork.

The future of AI gadgets is phones; Saudi Arabia and the UAE race to become global AI superpowers; Cohere looks towards the enterprise for success; most global companies are unprepared for AI.

Alexandru Voica's avatar
Alexandru Voica
Apr 26, 2024
∙ Paid

Share this post

Computerspeak by Alexandru Voica
Computerspeak by Alexandru Voica
Open models are great but they have a big catch; Synthesia releases Expressive Avatars; generative AI arrives in healthcare; AI assistants and busywork.
1
Share

The generative AI race welcomed another important contender last week when Meta released two pre-trained and fine-tuned Llama 3 language models with 8B and 70B parameters, respectively. Just in the past three months, Microsoft, Google, Databricks, Cohere and Mistral have all made foundational models available for broad use, creating a growing trend among technology companies.

First, let’s make one thing clear: most of these models were released in the form of open weights coupled with inference code, not open source code and datasets. The difference is important because code represents the well-structured, human-readable text written in a programming language that describes computer systems. Weights are just the output of training runs of a model on data and are almost impossible for a person to understand (you can think of them as very similar to the binaries of traditional software packages). There’s a separate debate between free software and open source software, but we won’t go there today.

Llama 3 has the best models existing today at the 8B and 70B parameter scale.

On its face, releasing the weights of AI models seems like an unequivocally good thing: researchers get access to state-of-the-art technology and can extend the models to create new variants of them. This allows smaller companies, researchers, and hackers to build new applications that wouldn’t have otherwise existed, as foundational models are incredibly data and compute intensive. It also creates an open ecosystem that spurs innovation in unexpected directions as different groups experiment with the models. For example, MBZUAI (the world’s first AI university) built BiMediX, a bilingual model for medical applications, by changing and fine-tuning Mistral’s Mixtral 8x7B mixture of experts model. The Yi-34B model from Chinese startup 01.ai uses Llama's architecture, reducing the effort required to build it from scratch and enabling the utilization of the same tools within the Llama ecosystem.

Open software also has a track record of producing more secure technologies and products, as more people can better stress test the code and discover bugs that wouldn’t have been found in closed environments.

However, the details of how large language models are created and the scale of resources required to build them make AI different to traditional software, raising questions about whether truly decentralized development is possible, even with an open source codebase.

There are three uncomfortable truths that advocates of openness in AI typically avoid.

First, despite opening their models, large tech companies are still likely to wield outsized influence over the trajectory of AI progress. The challenge stems from “the dirty secret of generative AI”: the code of foundational models is becoming more and more irrelevant. What’s more important is the immense computing power and data required to train the foundation models that can then be adapted for downstream applications. Only a handful of companies have the specialized AI supercomputers and training data measured in the trillions of words to create these large models.

So while a company may release the code or weights for an AI model under a permissive license, the ability to materially improve, update, or build new models at that scale will soon be limited only to major tech players. The open source license simply makes it easier for others to repurpose and adapt what has already been created by the company that developed the original foundation.

In addition, the comprehensive training datasets required to create large language models often include web crawled data, books, and databases to which the tech companies have preferential access or have spent years gathering and curating. This data disadvantage for those outside big tech puts up another barrier to entry.

As a result, we could see an AI development ecosystem where a small group of big tech companies open up their models, but still largely control the rate and direction of fundamental progress due to their unmatched resources for retraining and updating. Smaller players and researchers could build innovative applications atop the open sourced foundations, but would remain reliant on the tech giants to advance the technology.

Theoretically, big tech companies could embrace more of an authentic open source approach, working with academics and smaller teams to federate some of the training data and compute resources. But the fiercely competitive nature of the market and the immense economies of scale at play make it unlikely that these companies will cede control over the core model development pipeline.

Secondly, an open weights approach is a way to rapidly commoditize foundational models. If the market is crowded, there’s no faster way to eliminate competition than making a great product free. And given the high costs of training large models, very few startups (and even medium to large sized companies) can justify the price tag of trying to go head-to-head at the model level, unless they’re building for very targeted applications.

And thirdly, AI is a powerful technology unlike anything else we’ve seen in traditional software. As such, it gives bad faith actors unprecedented power to create harmful content at scale. While some debate abstract existential risks, real people are already experiencing the ramifications of unchecked AI today, with explicit deepfakes affecting everyone from high profile celebrities to primary school children. Given the lack of worldwide legislation on AI-related harms and the struggles of internet platforms to moderate AI-generated content, is it wise to continue making this technology openly available to everyone, including potential abusers?

Therefore, as AI capabilities rapidly advance, surface-level openness is unlikely to be enough. Instead, we should encourage real decentralization of power if we are serious about mitigating the risks around the control of such a powerfully transformative technology. For all the real benefits to opening up large AI models, we shouldn't be lulled into thinking the step of opening up alone will prevent the biggest tech firms from controlling the overarching direction and pace of AI progress.

And now, here are this week’s news:

❤️Computer loves

Our top news picks for the week - your essential reading from the world of AI

  • MIT Technology Review: An AI startup made a hyperrealistic deepfake of me that’s so good it’s scary

  • FT: AI service providers mine a rich seam of demand for efficiencies

  • New York Times: Generative A.I. Arrives in the Gene Editing World of CRISPR

  • The Verge: The future of AI gadgets is just phones

  • CNBC: Less burnout for doctors, better clinical trials, among the benefits of AI in health care

  • New York Times: ‘To the Future’: Saudi Arabia Spends Big to Become an A.I. Superpower

  • Bloomberg: Most Global Tech Leaders See Their Companies Unprepared for AI

  • Fortune: The CEO of OpenAI rival Cohere shakes off the haters: ‘We’re still sort of the underdog’

  • Bloomberg: Caught Between the US and China, a Powerful AI Upstart Chooses Sides

  • Axios: AI shakes up corporate boards

  • WSJ: At Moderna, OpenAI’s GPTs Are Changing Almost Everything

  • New York Times: How a Virtual Assistant Taught Me to Appreciate Busywork

  • Washington Post: The AI hype bubble is deflating. Now comes the hard part.

⚙️Computer does

AI in the wild: how artificial intelligence is used across industry, from the internet, social media, and retail to transportation, healthcare, banking, and more

  • Fortune: Cargill leans on regenerative agriculture and generative AI to feed the planet

  • TechCrunch: Rabbit’s R1 is a little AI gadget that grows on you

  • MIT Technology Review: This creamy vegan cheese was made with AI

  • BBC: How robots are taking over warehouse work

  • The Verge: The Ray-Ban Meta Smart Glasses have multimodal AI now

  • TechCrunch: Sanctuary’s new humanoid robot learns faster and costs less

  • Business Insider: AI is helping Amazon send fewer small items in comically large boxes

  • TechCrunch: How United Airlines uses AI to make flying the friendly skies a bit easier

  • Fortune: AI is shaking up how sports like rugby, soccer and cricket are played—and could mint big money for sports clubs

  • AP: A coffee roastery in Finland has launched an AI-generated blend. The results were surprising

🧑‍🎓Computer learns

Interesting trends and developments from various AI fields, companies and people

  • Axios: Space industry races to put AI in orbit

  • WSJ: Microsoft CEO to Visit Southeast Asia With AI on Agenda

  • VentureBeat: Estée Lauder and Microsoft partner to help beauty brands use generative AI

  • Reuters: Beijing city to subsidise domestic AI chips, targets self-reliance by 2027

  • Business Insider: Google's 'fastest-growing' moonshot that spun out into an AI earbuds startup aims to ship its first product this year

  • VentureBeat: Visa’s CIO shares insights on generative AI’s transformative potential in payments industry

  • Axios: OpenAI's Chris Lehane says AI is "critical infrastructure"

  • Time: Tech CEOs Say Ethical A.I. and Innovation Are ‘Two Sides of the Same Coin’

  • Wired: Meta’s Open Source Llama 3 Is Already Nipping at OpenAI’s Heels

  • ZDNet: How AI hallucinations could help create life-saving antibiotics

  • The Verge: Samsung and Google tease “exciting things” for AI on Android and Galaxy hardware.

  • Bloomberg: Inside the China-US Competition for AI Experts

  • VentureBeat: Salesforce Einstein Copilot brings new reasoning and actions to enterprise generative AI 

  • VentureBeat: Nvidia CEO Jensen Huang personally delivers first DGX H200 to OpenAI

  • FT: AI could kill off most call centres, says Tata Consultancy Services head

  • Reuters: Toyota and Nissan pair up with Tencent and Baidu for China AI arms race

  • MIT Technology Review: Chatbot answers are all made up. This new tool helps you figure out which ones to trust.

  • The Atlantic: Would Limitlessness Make Us Better Writers?

  • The Verge: Apple’s new AI model hints at how AI could come to the iPhone

  • WSJ: Amazon Introduces Custom AI Capabilities in Race Against Cloud Rivals

  • VentureBeat: Snowflake launches Arctic, an open ‘mixture-of-experts’ LLM to take on DBRX, Llama 3

  • Axios: Generative AI is still a solution in search of a problem

  • TechCrunch: Why code-testing startup Nova AI uses open source LLMs more than OpenAI

  • The Verge: A morning with the Rabbit R1: a fun, funky, unfinished AI gadget

  • WSJ: Why the AI Industry’s Thirst for New Data Centers Can’t Be Satisfied

  • VentureBeat: DeepMind researchers discover impressive learning capabilities in long-context LLMs

  • The Verge: Adobe’s impressive AI upscaling project makes blurry videos look HD

  • Fortune: Move over, Black Mirror; the BBC is showing AI can be a force for good in media

  • Reuters: Tesla could start selling Optimus robots by the end of next year, Musk says

  • Sifted: AI could save lives in the NHS – but a lack of adoption pathways are holding it back

  • New York Times: Meta’s A.I. Assistant Is Fun to Use, but It Can’t Be Trusted

  • MIT Technology Review: Job titles of the future: AI prompt engineer

  • Business Insider: Sundar Pichai admits the generative AI boom took Google by surprise

  • BBC: Essex University's AI brain study brings 'hope' to childhood trauma survivors

  • Bloomberg: Inside Palantir’s AI Sales Secret Weapon: Software Boot Camp

  • Reuters: Adobe to bring full AI image generation to Photoshop this year

  • WSJ: Salesforce Calls for AI Emissions Regulations as Concerns Grow Over Tech Sector’s Carbon Footprint

  • FT: BCG says AI consulting will supply 20% of revenues this year

  • MIT Technology Review: Three things we learned about AI from EmTech Digital London

  • The Verge: Netflix’s Atlas is looking more and more like a buddy cop movie about working with AI

  • VentureBeat: OpenAI shrugs off Meta's Llama 3 ascent with new enterprise AI features

  • ZDNet: Vyond's video generator adds AI that businesses will love. Try it for yourself

  • Reuters: AI boom to fuel natural gas demand in coming years, report says

  • VentureBeat: Mabl launches an AI-powered low-code mobile app testing service

  • Reuters: Coca-Cola signs $1.1 bln deal to use Microsoft cloud, AI services

  • VentureBeat: Tredence’s GenAI platform attracts top clients, fuels enterprise AI adoption

  • VentureBeat: Amazon Bedrock continues to lay down generative AI foundation for the cloud

  • VentureBeat: Adobe launches Firefly 3 with full AI image generation in Photoshop

  • Reuters: Vietnam's FPT to invest $200 mln in AI factory using Nvidia chips

  • VentureBeat: More than half of U.S. has tried generative AI according to Adobe Analytics

  • Reuters: UAE-based AI firm G42 announces collaboration with U.S. group Qualcomm

  • Business Insider: As most AI execs scramble for more data, Mark Zuckerberg says there's actually something more 'valuable'

  • Reuters: Microsoft launches lightweight AI model

  • VentureBeat: Alethea AI launches emotive and expressive AI avatars on Coinbase blockchain

  • Axios: AI computer on your ears

  • TechCrunch: Why vector databases are having a moment as the AI hype cycle peaks

  • Bloomberg: An AI Star Seeks to Bring Self-Driving Cars to Japan by 2030

  • VentureBeat: Why LLMs are predicting the future of compliance and risk management

  • The Verge: Microsoft hires former Meta exec to bolster AI supercomputing team

  • VentureBeat: Groq’s breakthrough AI chip achieves blistering 800 tokens per second on Meta’s LLaMA 3

  • Fortune: There’s an early winner in the race to feed AI’s infrastructure demands: Private equity

  • FT: US seeks alliance with Abu Dhabi on artificial intelligence

  • Business Insider: Mark Zuckerberg did not see the GenAI wave coming. He prepared anyway.

  • Business Insider: Former OpenAI leader has an easy trick for surviving AI in the workplace

  • The Guardian: BBC to invest in AI to help transform its education services

  • The Guardian: Where do we draw the line on using AI in TV and film?

  • The Guardian: A revolution for sport? Olympic vision for AI innovations laid out by IOC

Keep reading with a 7-day free trial

Subscribe to Computerspeak by Alexandru Voica to keep reading this post and get 7 days of free access to the full post archives.

Already a paid subscriber? Sign in
© 2025 Alexandru Voica
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share