What Is AI?

Artificial Intelligence (AI) is a technology that empowers computers and machines to mimic human abilities such as learning, understanding, problem-solving, decision-making, creativity, and independent action

Where is AI used?

Organisations use AI to help drive innovation, empower their teams, and streamline operations in variety of ways. Depending on how it is implemented, AI capabilities are often bolstered with predictive analytics, machine learning (ML), and other functionalities. AI use cases include but are not limited to:

  • Leveraging AI data centres to train AI models
  • Assisting customers with the help of chatbots
  • Streamlining supply chain management processes
  • Automating repetitive tasks to free up resources and teams
  • Assisting with ideation and creative brainstorming
  • Reducing query response times
  • Accelerating decision-making processes
  • Reducing costs via AI-assisted optimisation and automation
  • Translating text, audio, and/or speech
  • Consolidating research findings and stats quickly

AI PCs

The hardware we use to drive innovation, streamline processes, and manage everyday operations is changing. Advanced architectures such as reduced instruction set computing (RISC) machines—commonly known as ARM development—and complex instruction set computing (CISC) architectures like x86 are both playing critical roles in the computing industry. With Apple, Microsoft, Broadcom, Intel, and other companies investing heavily into AI-enabling technologies, we have entered the age of AI PCs. These systems are optimised to handle a wide range of AI-enabled tasks, including but not limited to voice recognition, natural language processing, and machine learning. AI-specific hardware accelerates many of these tasks on-device, allowing powerful AI inference and even training in everyday machines.

AI data centres

To power deep learning and train AI models, organisations are leveraging the performance and expanded throughput offered by AI data centres. These are facilities that are home to large quantities of hardware including graphical processing units (GPUs) and AI acceleration systems. As a recent Forbes article exploring their capabilities notes, these deliver “substantial computational power," collectively consuming massive amounts of power and even requiring state-of-the-art cooling solutions.

What is SOC AI?

Security operations centres (SOCs) can leverage AI to more efficiently allocate resources and mitigate risk. Through deep learning, automation, and other capabilities, they can accelerate their risk identification and response measures, particularly if utilising a cybersecurity platform that consolidates solutions and integrates AI to streamline their operations further.

AI security risks

With tools including ChatGPT, OpenAI, and Microsoft Copilot being easily accessible, threat actors are continually attempting to access sensitive data. In some cases, their aim is to target AI tools to manipulate their behavior to operate against its intended use. Key AI security risks include rogue AI, fraud automation, and insufficient data governance.

Organisations must not only keep pace but get ahead of cybercriminals by ensuring risk-aware and compliant adoption of AI technology. Developing a deeper understanding of AI security risks is a vital part of this strategy.

AI company policies

A risk-aware policy that provides guidance on correct AI use is an important point of reference for employees. Ensuring it is followed and kept up to date will help minimise the risk levied against your organisation. Having the right policies and procedures in place is essential to be compliant and maintain effective data security. Exploring examples from federal and industry regulators and working with peers can help inform the drafting of your own AI policy.

Deepfakes

As generative AI (GenAI) technology continues to advance, deep fakes are being made increasingly convincing. With threat actors using them to manipulate individuals or groups into believing the image, video, or text generated to be authentic and trustworthy, they pose a substantial data security risk. Whether or not AI plays a role, the intention of cybercriminals in using either approach—to mislead, steal, and/or defraud—remains the same.

How to secure AI

Understanding how AI implementations function—including how they leverage and potentially retain data—helps to inform an effective cybersecurity response. As organisations continue to imagine and innovate with AI, malicious actors are adapting accordingly to take advantage of vulnerabilities. With the threat landscape constantly evolving in tandem with AI itself, organisations should strive to proactively secure their AI implementations.

Recommended actions from the Open Web Application Security Project (OWASP) overview on AI security

  • Implement AI governance
  • Extend your security practices via OWASP assets and controls (as well as AI security solutions)
  • Ensure that any AI suppliers you use have implemented appropriate security controls
  • Limit the impact of your AI implementation by minimising data and privileges
  • Prevent runaway use and denial of service with rate-limiting
  • Implement guardrails in the form of oversight processes (including human oversight)

In addition, if developing your own AI systems and regardless of whether you train your own models, OWASP recommends the following:

  • Involve your data and AI engineering into your secure software development practices
  • Apply appropriate process and technical controls (through learning about AI threats)

Read the OWASP AI security overview for additional details and technical insights.

Machine learning (ML) data security

With GenAI leveraging ML capabilities for data analysis and creative output, new risks are emerging. “Machine learning data security must also consider data integrity in transit and during processing,” notes a Global Cybersecurity Alliance (GCA) article on ML data security. “Compromised data integrity can skew model outputs. It can lead to inaccurate or biased decisions with potentially far-reaching consequences.”

Proactive steps are explored in detail within this article:

  • Access control and authentication
  • Regular data audits and monitoring
  • Data anonymisation and pseudonymisation
  • Secure sharing practices

What is an AI model?

AI models are structures made up of an architecture and parameter values that allow the system to perform tasks like making predictions or generating content, which is called inference. These tasks can range from answering queries, detecting patterns in data, recognising behaviors, and more. AI models typically go through a training process to learn the best parameter values for effective inference.

Depending on your organisation’s needs, goals, compliance requirements, and budget—among other factors—a wide range of ideal AI models may be under consideration for implementation. However, it’s important to note that every AI model has its own inherent level of risk, and there are also different types of AI models to consider.

Types of AI models

Narrow/traditional AI

Much of the most widely implemented and established AI technology we have today is referred to as traditional or narrow AI. While it can adapt to user queries and/or needs, it can only perform predetermined tasks, often within one domain of expertise. Examples of narrow AI include text-based chatbots in customer support portals, virtual assistants such as Siri or Google Assistant, and language detection software with auto-translate capabilities.

According to IBM’s Data and AI Team, there are two functional categories of narrow AI:

Reactive machine AI

As the term suggests, reactive machine AI can only make use of the information that you feed it in the present moment. It can actively engage with its environment and users but, unable to memorise what it receives, it cannot self-improve. Content recommendations built into streaming and social media platforms make use of reactive machine AI, as do tools designed to perform predictive analyses of real-time data.

Limited memory AI

Limited memory AI leverages past and presently available data to better assist you. The “limited” distinction refers to it being unable to hold onto your provided data indefinitely, essentially relying on short-term memory. The data that it can access, however, is leveraged to help continually optimise its performance and capabilities. In other words, its environment and your input help to train it on how best to respond. Virtual assistants fall under this category, for instance.

Frontier AI

While narrow AI is used in a variety of implementations, frontier AI models—more commonly referred to as GenAI—are also receiving plenty of attention and investment. These are even more advanced, experimental, and future facing AI models by design. As the term implies, GenAI is designed to generate content, either from prompt inputs or accessing existing data. Standout examples include GPT-4 and Google Gemini Ultra.

The Artificial Intelligence Index Report 2024 by Stanford University estimates that frontier AI training costs have reached “unprecedented levels,” with Google Gemini Ultra alone costing US $191 million. In addition, it states that industry is a significant driver of frontier AI research, producing 51 “notable machine learning models” in 2023 compared to 15 in academia. Yet, at the same time, 21 such models emerged from industry-academia collaborations. The report also notes that, despite declining private investment in 2022, GenAI funding has surged to US $25.2 billion, and “all major players […] reported substantial fundraising rounds.” 

Summarising AI model types

“Traditional AI excels at pattern recognition, while generative AI excels at pattern creation. Traditional AI can analyse data and tell you what it sees, but generative AI can use that same data to create something entirely new,” author Bernard Marr summarises in The Difference Between Generative AI and Traditional AI: An Easy Explanation for Everyone (Forbes). “Both generative AI and traditional AI have significant roles to play in shaping our future, each unlocking unique possibilities. Embracing these advanced technologies will be key for businesses and individuals looking to stay ahead of the curve in our rapidly evolving digital landscape.”

Additional AI-associated terms

Algorithm

A set of step-by-step instructions designed to solve a problem or perform a task. It defines a sequence of operations that can be executed by a computer.

Deep learning

An ML subset where algorithms, inspired by the structure and function of the human brain's neural networks, learn from large amounts of data. ‘Deep’ refers to the large number of layers in which these artificial neurons are organised. Deep learning excels in tasks like image and speech recognition, natural language processing, and more complex pattern recognition.

AI agent

A system designed to perceive its environment and take actions to maximise its chances of achieving specific goals. It uses sensors to gather information and algorithms to make decisions, take actions and evaluate the effect, often learning and adapting over time.

Synthetic media

This is content produced or manipulated using AI techniques, such as deep learning. It includes generated images, videos, and audio that convincingly simulate real-world elements, blurring the line between authenticity and simulation.

Large language model (LLM)

LLMs refer to AI models with billions of parameters, such as GPT-4, that are trained on vast datasets to manipulate and generate human-like text. This enables various language-related tasks and applications. Transformers are currently the dominant architecture for LLMs.

Foundation model

This is usually a deep learning model trained on a broad data set, which can then be repurposed for many different tasks. LLMs are examples of foundation models, capable of being specialised for language, code, images, audio, a combination of modalities, or they can be multi-modal. Foundation models can also be finetuned for specialised applications, like chatbots.

Related Research

Related Articles