OpenAI and Figma have launched an integration that allows teams to switch between code and design faster – without extra tools or manual synchronization.
Kubetorch has joined the PyTorch ecosystem, simplifying the process of running ML tasks on Kubernetes by abstracting complex infrastructure behind simple Python code.
Helion, a DSL for writing fast ML kernels, has gained a new automatic tuning mechanism based on Bayesian optimization that saves developers' time.
AI: Events
How to Train Large Language Models Without Constantly Babysitting the Terminal
Technical context • Infrastructure
AMD demonstrates how to set up LLM training on GPU clusters so that failures are handled automatically, eliminating the need for manual intervention.
AI: Events
How AMD Is Teaching Neural Networks to Work Together: Ray and ROCm 7 for Large-Scale ML Tasks
Technical context • Infrastructure
AMD has explained how to run distributed ML tasks on GPUs using Ray and ROCm 7 – from model training to creating agent-based systems.
AMD has released JAX-AITER, a library of pre-built, optimized computational blocks for developing large AI models on AMD GPUs using the JAX framework.