Artificial intelligence technology is developing at high speed and is transforming many aspects of modern life.
However, some experts fear that it could be used for malicious purposes and may threaten jobs.
What is AI and how does it work?
AI allows a computer to think, act, and respond almost as if it was a human.
Computers can be fed huge amounts of information and trained to identify the patterns in it, in order to make predictions, solve problems, and even learn from their own mistakes.
As well as data, AI relies on algorithms – lists of rules which must be followed in the correct order to complete a task.
The technology is behind the voice-controlled virtual assistants Siri and Alexa. It lets Spotify, YouTube and BBC iPlayer suggest what you might want to play next, and helps Facebook and Twitter decide which social media posts to show users.
AI lets Amazon analyse customers’ buying habits to recommend future purchases – and the firm is also using the technology to crack down on fake reviews.
Two powerful AI-driven applications or apps which have become very high profile in recent months are ChatGPT and Snapchat My AI.
They are examples of what is called “generative” AI.
This uses the patterns and structures it identifies in vast quantities of source data to generate new and original content which feels like it has been created by a human.
The AI is coupled with a computer programme known as a chatbot, which “talks” to human users via text.
The apps can answer questions, tell stories and write computer code.
But both programmes sometimes generate incorrect answers for users, and can reproduce the bias contained in their source material, such as sexism or racism.
With few rules currently in place governing how AI is used, experts have warned that its rapid growth could be dangerous. Some have even said AI research should be halted.
In May, Geoffrey Hinton, widely considered to be one of the godfathers of artificial intelligence, quit his job at Google, warning that AI chatbots could soon be more intelligent than humans.
Later that month, the US-based Center for AI Safety published a statement supported by dozens of leading tech specialists.
They argue AI could be used to generate misinformation that could destabilise society. In the worst-case scenario, they say machines might become so intelligent that they take over, leading to the extinction of humanity.
The EU’s competition chief Margrethe Vestager says “guardrails” are needed to counter the biggest risks posted by AI.
In particular she is concerned about the role AI could play in making decisions that affect people’s livelihoods such as loan applications, adding there was “definitely a risk” that AI could be used to influence elections.
Members of the European Parliament voted in favour of the EU’s proposed Artificial Intelligence Act, which will put in place a strict legal framework for AI, which companies would need to follow.
The legislation – which is expected to come into effect in 2025 – categorises applications of AI into levels of risk to consumers, with AI-enabled video games or spam filters falling into the lowest risk category.

