Understanding AI

Recent advances in AI software are powering new opportunities and experiences across industries. At the same time, AI is creating new challenges for IT professionals.

Red Hat Enterprise Linux AI

Develop, test, and deploy large language models (LLMs) with optimized inference capabilities.

What is AI?

Artificial intelligence (AI) generally refers to computer science processes and statistical algorithms that are able to simulate and augment human intelligence.

In other words, AI describes systems capable of acquiring knowledge and applying insights to enable problem solving.

Read more about specific AI topics

Generative AI

Generative AI relies on deep learning models to create new content.

Machine learning

Machine learning (ML) trains a computer to find patterns and make predictions.

Deep learning

Deep learning processes data using an algorithm inspired by the human brain.

Foundation models

A foundation model is a type of ML model pretrained to perform a range of tasks.

Large language models

Large language models (LLMs) understand and generate human language.

AI infrastructure

AI infrastructure helps data scientists and developers work efficiently.

MLOps

MLOps integrates ML models into software development processes.

AIOps

AIOps use AI to enhance or partially replace a range of IT operations processes and tasks.

AI/ML use cases

Different industries benefit from different applications of AI/ML.

AI in healthcare

Across the healthcare market, AI can serve life sciences, providers, and payers.

AI in telecommunications

Service providers use AI to overcome network complexity, operational inefficiencies, and market competition.

AI in banking

AI can improve customer service and operational efficiency in banking.

Edge AI

Edge AI lets devices make smarter decisions faster.

LLMOPs

LLMOps manages and automates the lifecycle of LLMs.

AI platforms

An AI platform is a collection of technologies to develop, train, and run machine learning models.

InstructLab

InstructLab simplifies the process of customizing LLMs with private data.

Retrieval-augmented generation

Retrieval-augmented generation (RAG) links external resources to an LLM for a more accurate output.

Intelligent applications

Intelligent applications use AI to augment a human workflow.

RAG vs. fine-tuning

Fine-tuning and RAG both aim to improve LLMs, but use different methods. RAG avoids altering the model, while fine-tuning requires adjusting its parameters.

Granite models

Granite is a series of LLMs created by IBM for enterprise applications that support gen AI use cases like language and code.

Agentic AI

Agentic AI is a software system designed to interact with data and tools in a way that requires minimal human intervention.

Predictive AI

Compare predictive AI and gen AI to find out which best fits your AI use cases.

LLMs vs. SLMs

LLMs and small language models (SLMs) are both types of AI systems trained to interpret human language, including programming languages.

AI inference

AI inference is when an AI model provides an answer based on data.

vLLM

vLLM is a library of open source code that helps large language models (LLMs) perform calculations more efficiently and at scale.

LoRA vs. QLoRA

LoRA and QLoRA both help fine-tune LLMs more efficiently, but manipulate the model and utilize storage differently.

Parameter-efficient fine-tuning

Parameter-efficient fine-tuning (PEFT) is a set of techniques that adjusts only a portion of parameters within an LLM to save resources.

Enterprise AI

AI at the enterprise comes with its own challenges and benefits on a larger scale.

Models-as-a-Service

MaaS is an approach to delivering AI models as shared resources, allowing users within an organization to access them on demand.

Agentic AI vs. generative AI

Learn how each technology works, their unique strengths, and how they can collaborate for smarter solutions.

AI security

AI security defends AI applications against malicious attacks that weaken workloads and manipulate data.

Model Context Protocol (MCP)

Model Context Protocol (MCP) is an open source protocol that enables 2-way connection and standardized communication between AI applications and external services.

llm-d

llm-d is an open source framework to speed up LLM inference at scale.

Distributed inference

Distributed inference lets AI models process workloads more efficiently by dividing the labor of inference across a group of interconnected devices.

Resources

E-book

Discover how organizations speed AI/ML adoption with Red Hat OpenShift

See how AI/ML can help make your business more competitive.

Checklist

Top 5 ways to implement MLOps successfully in your organization

Discover how Red HatOpenShift can help on your MLOps journey.

E-book

The adaptable enterprise: Why AI readiness is disruption readiness

Learn how Red Hat's open source AI solutions offer a resilient foundation for innovation.

Success stories

Banco Galicia speeds new customer onboarding

Using an AI-based natural language processing (NLP) solution Red Hat OpenShift, the bank cut verification times from days to minutes with 90% accuracy.​

EJIE

Eusko Jaularitzaren Informatika Elkartea (EJIE)

To help preserve the Basque language, EJIE created the Itzuli application, an AI-based translation application to help residents translate Basque to and from Spanish, French, and English.

Boston University builds an educational platform

Using Red Hat OpenShift AI, Boston University scaled a learning environment to accommodate hundreds of computer science and computer engineering users.

More about AI/ML

Products

A foundation model platform used to seamlessly develop, test, and run Granite family LLMs for enterprise applications.

An AI-focused portfolio that provides tools to train, tune, serve, monitor, and manage AI/ML experiments and models on Red Hat OpenShift.

An enterprise application platform with a unified set of tested services for bringing apps to market on your choice of infrastructure.

Red Hat Ansible Lightspeed is a generative AI service designed by and for Ansible automators, operators, and developers.

Resources

e-book

Top considerations for building a production-ready AI/ML environment

Webinar

Getting the most out of AI with open source and Kubernetes