Sydney: In the world of artificial intelligence (AI), a battle is underway. On one side are companies that believe in keeping the datasets and algorithms behind their advanced software private and confidential. On the other are companies that believe in allowing the public to see what's under the hood of their sophisticated AI models. Think of this as the battle between open- and closed-source AI.
In recent weeks, Meta, the parent company of Facebook, took up the fight for open-source AI in a big way by releasing a new collection of large AI models. These include a model named Llama 3.1 405B, which Meta's founder and chief executive, Mark Zuckerberg, says is the first frontier-level open-source AI model.
For anyone who cares about a future in which everybody can access the benefits of AI, this is good news. The danger of closed-source AI and the promise of open-source AI
Closed-source AI refers to models, datasets and algorithms that are proprietary and kept confidential. Examples include ChatGPT, Google's Gemini and Anthropic's Claude.
Though anyone can use these products, there is no way to find out what dataset and source codes have been used to build the AI model or tool. While this is a great way for companies to protect their intellectual property and their profits, it risks undermining public trust and accountability. Making AI technology closed-source also slows down innovation and makes a company or other users dependent on a single platform for their AI needs. This is because the platform that owns the model controls changes, licensing and updates.
There are a range of ethical frameworks that seek to improve the fairness, accountability, transparency, privacy and human oversight of AI. However, these principles are often not fully achieved with closed-source AI due to the inherent lack of transparency and external accountability associated with proprietary systems. In the case of ChatGPT, its parent company, OpenAI, releases neither the dataset nor code of its latest AI tools to the public. This makes it impossible for regulators to audit it. And while access to the service is free, concerns remain about how users' data are stored and used for retraining models.
By contrast, the code and dataset behind open-source AI models are available for everyone to see. This fosters rapid development through community collaboration and enables the involvement of smaller organisations and even individuals in AI development. It also makes a huge difference for small and medium size enterprises as the cost of training large AI models is colossal.
Perhaps most importantly, open-source AI allows for scrutiny and identification of potential biases and vulnerabilities. However, open-source AI does create new risks and ethical concerns. For example, quality control in open-source products is usually low. As hackers can also access the code and data, the models are also more prone to cyberattacks and can be tailored and customised for malicious purposes, such as retraining the model with data from the dark web.
An open-source AI pioneer
Among all leading AI companies, Meta has emerged as a pioneer of open-source AI. With its new suite of AI models, it is doing what OpenAI promised to do when it launched in December 2015 namely, advancing digital intelligence in the way that is most likely to benefit humanity as a whole, as OpenAI said back then.