订阅内容

I’m excited to announce the general availability of Red Hat Enterprise Linux AI (RHEL AI) 1.2, our generative AI (gen AI) foundation model platform to develop, test and run large language models (LLMs) for enterprise applications. RHEL AI combines open source Granite LLMs with InstructLab model alignment tools on a purpose-built RHEL image optimized for gen AI use cases.

Building on the success of our initial release of RHEL AI 1.1 on September 5, 2024, this version furthers our commitment to empowering developers, AI engineers and data scientists by lowering the barriers of entry and adoption to enterprise AI. RHEL AI 1.2 brings significant enhancements, allowing organizations to more efficiently fine-tune and deploy LLMs using private, confidential and sovereign data to better align to enterprise use cases. These improvements, powered by InstructLab and a comprehensive software stack, now support a wider range infrastructure options, including NVIDIA accelerated computing and software and the newly introduced AMD Instinct accelerators. We intend to continue expanding our hardware accelerator support with partners like Intel in upcoming releases.

Key highlights of RHEL AI 1.2:

  • Support for Lenovo ThinkSystem SR675 V3 servers
    RHEL AI 1.2 is now supported on Lenovo ThinkSystem SR675 V3 servers wit NVIDIA accelerated computing. Users can also take advantage of factory preload options for RHEL AI on these servers, making deployment faster and easier.

  • Support for AMD Instinct Accelerators (technology preview)
    Language models require powerful computing resources, and RHEL AI now supports AMD Instinct Accelerators with the full ROCm software stack, including drivers, libraries and runtimes. With RHEL AI 1.2, organizations can leverage AMD Instinct MI300x GPUs for both training and inference, and AMD Instinct MI210 GPUs for inference tasks.

  • Availability on Azure and GCP: 
    RHEL AI is now available on Azure and Google Cloud Platform (GCP). With this users will be able to download RHEL AI from Red Hat and bring them to Azure and GCP and create RHEL AI based GPU instances.
  • Training checkpoint and resume
    Long training runs during model fine tuning can now be saved at regular intervals, thanks to periodic checkpointing. This feature allows InstructLab users to resume training from the last saved checkpoint instead of starting over, saving valuable time and computational resources.

  • Auto-Detection of hardware accelerators
    The ilab CLI can now automatically detect the type of hardware accelerator in use and configure the InstructLab pipeline accordingly for optimal performance, reducing the manual setup required.

  • Enhanced training with PyTorch FSDP (technology preview)
    For multi-phase training of models with synthetic data, ilab train now uses PyTorch Fully Sharded Data Parallel (FSDP). This dramatically reduces training times by sharding a model’s parameters, gradients and optimizer states across data parallel workers (e.g., GPUs). Users can pick FSDP for their distributed training by using ilab config edit.

These are just a few of the exciting new features in RHEL AI 1.2. Many more improvements and bug fixes are included, making this release a powerful tool for AI development.

Don’t miss out on these powerful new features! Download RHEL AI 1.2 today and deploy it on-premises or across all major public cloud providers and take your AI development to the next level.

Important notice:

With the introduction of RHEL AI 1.2, we will be deprecating support for RHEL AI 1.1 in 30 days. Please ensure your systems are upgraded to RHEL AI 1.2 to continue receiving support.


关于作者

UI_Icon-Red_Hat-Close-A-Black-RGB

按频道浏览

automation icon

自动化

有关技术、团队和环境 IT 自动化的最新信息

AI icon

人工智能

平台更新使客户可以在任何地方运行人工智能工作负载

open hybrid cloud icon

开放混合云

了解我们如何利用混合云构建更灵活的未来

security icon

安全防护

有关我们如何跨环境和技术减少风险的最新信息

edge icon

边缘计算

简化边缘运维的平台更新

Infrastructure icon

基础架构

全球领先企业 Linux 平台的最新动态

application development icon

应用领域

我们针对最严峻的应用挑战的解决方案

Original series icon

原创节目

关于企业技术领域的创客和领导者们有趣的故事