AI ethics or AI Law concept. Developing AI codes of ethics. Compliance, regulation, standard , business policy and responsibility for guarding against unintended bias in machine learning algorithms.
Cover AI’s power consolidation raises pressing ethical questions (Photo: Getty Images)
AI ethics or AI Law concept. Developing AI codes of ethics. Compliance, regulation, standard , business policy and responsibility for guarding against unintended bias in machine learning algorithms.

AI development is currently held by a small number of companies. Public vigilance can help ensure they stick to ethical use of the technology

Warren Buffett got it partly right about AI. The billionaire investor and philanthropist told CNN earlier this year: “We let a genie out of the bottle when we developed nuclear weapons... AI is somewhat similar—it’s part way out of the bottle.”

Buffett’s rationale is that, much like nuclear weapons, AI holds the potential to unleash profound consequences on a vast scale, both for better or worse. Like nuclear weapons, AI is concentrated in the hands of the few. In AI’s case, tech companies and nations. This is a comparison that is not often talked about.

As these companies push the boundaries of innovation, a critical question emerges: Are we sacrificing fairness and societal well-being on the altar of progress?

Read more: How should we regulate AI?

One study suggests that Big Tech’s influence is ubiquitous across all streams of the policy process, reinforcing their position as “super policy entrepreneurs”. This allows them to steer policies to favour their interests, often at the expense of broader societal concerns. This concentrated power also allows these corporations to mould AI technologies using vast datasets reflective of specific demographics and behaviours, often at the expense of broader society.

The result is a technological landscape that, while rapidly advancing, may inadvertently deepen societal divides and perpetuate existing biases.

Ethical concerns

Tatler Asia
Photo: Getty Images
Above A widely used healthcare algorithm underestimated Black patients’ needs, highlighting how AI systems trained on biased data can exacerbate existing disparities (Photo: Getty Images)
Photo: Getty Images

The ethical concerns stemming from this concentration of power are significant. If an AI model is primarily trained on data reflecting one demographic’s behaviour, it may perform poorly when interacting with or making decisions about other demographics, potentially leading to discrimination and social injustice.

This bias amplification is not just a theoretical concern but a pressing reality that demands immediate attention. Porcha Woodruff, for example, a pregnant Black woman, found herself wrongfully arrested due to a facial recognition error—a stark reminder of AI’s real-world consequences.

In healthcare, a widely used algorithm severely underestimated Black patients’ needs, leading to inadequate care and perpetuating existing disparities. These cases underscore a troubling pattern: AI systems, trained on biased data, amplify societal inequalities.

Consider the algorithms driving these AI systems, developed mainly within environments that lack sufficient oversight regarding fairness and inclusivity.

Read more: Doctor Anywhere founder Lim Wai Mun on creating a blueprint for a sustainable tech‑enabled healthcare company

Developing bias

Tatler Asia
Photo: Getty Images
Above AI applications developed by profit-driven corporations risk perpetuating biases against underrepresented communities (Photo: Getty Images)
Photo: Getty Images

Consequently, AI applications in areas such as facial recognition, hiring practices and loan approvals might develop biased outcomes, affecting underrepresented communities disproportionately.

This risk is accentuated by the business model of these corporations, which emphasises rapid development and deployment over rigorous ethical review, putting profits above proper consideration of the long-term societal impacts.

To counter these challenges, a change in AI development is urgently needed.

Broadening the influence beyond big tech companies to include independent researchers, ethicists, public interest groups and government regulators working collaboratively to establish guidelines which prioritise ethical considerations and societal well-being in AI development would be a good start. Governments have a pivotal role to play.

Read more: Who is Jensen Huang, the Nvidia tech billionaire leading the global AI revolution?

Tatler Asia
Public vigilance and participation are essential for holding companies and governments accountable (Photo: Getty Images)
Above Public vigilance and participation are essential for holding companies and governments accountable (Photo: Getty Images)
Public vigilance and participation are essential for holding companies and governments accountable (Photo: Getty Images)

Stringent antitrust enforcement would limit big tech’s power and promote competition. An independent watchdog with authority to sanction Big Tech practices would also help along with increasing public participation in policymaking and requiring transparency in tech companies’ algorithms and data practices.

Global cooperation on fostering ethical standards and investments in educational programmes to empower citizens to understand the impact of technology on society will further support these efforts.

The academic world, too, can step up. Researchers can advance methods to detect and neutralise biases in AI algorithms and training data. By engaging the public, academia can ensure diverse voices are heard in the shaping of AI policy.

Read more: Why humans still have the upper hand over AI—for now

Public vigilance and participation are indispensable for holding companies and governments accountable. The public can exert market pressure by choosing AI products from companies that demonstrate ethical practices.

While regulating AI would help prevent the concentration of its power among the few, antitrust measures which curb monopolistic behaviour, promote open standards, and support smaller firms and startups could help steer AI advancements towards the public good.

A unique opportunity

Tatler Asia
Photo: Getty Images
Above Democratising access to AI tools enables researchers, startups and educational institutions to engage with advanced tools and drive innovation across diverse sectors (Photo: Getty Images)
Photo: Getty Images

Nonetheless, the challenge remains that developing AI requires substantial data and computational resources, which can be a significant hurdle for smaller players. This is where open-source AI presents a unique opportunity to democratise access, potentially creating more innovation across diverse sectors.

Allowing researchers, startups, and educational institutions equal access to engage with state-of-the-art AI tools levels the playing field.

The future of AI is not predetermined. Taking action now can shape a technological landscape that reflects our collective values and aspirations, ensuring the benefits of AI are shared equitably across society. The question is not whether we can afford to take these steps but whether we can afford not to.


Arif Perdana is an Associate Professor at Monash University Indonesia, specialising in digital strategy, and data science. He is the director of Action Lab, Indonesia.

Ridoan Karim is a Lecturer in Business Law, and Deputy Director of Undergraduate Studies at the School of Business, Monash University Malaysia. He has taught and researched in the fields of business and cyber law.

Originally published under Creative Commons by 360info

Topics