Home Venture/Startup Meet Felafax: An AI Startup Building an Open-Source AI Platform for Next-Generation AI Hardware, Reducing Machine Learning ML Training Costs by 30%

Meet Felafax: An AI Startup Building an Open-Source AI Platform for Next-Generation AI Hardware, Reducing Machine Learning ML Training Costs by 30%

by WeeklyAINews
0 comment

It’s a problem to spin up AI workloads on the cloud. The prolonged coaching course of entails putting in a number of low-level dependencies, which could result in notorious CUDA failures. It additionally consists of attaching persistent storage, ready for the system besides up for 20 minutes, and far more. Machine studying (ML) assist for GPUs that aren’t NVIDIA is missing. Alternatively, Google TPUs and different various chipsets have a 30% decrease complete price of possession whereas nonetheless offering superior efficiency. The growing measurement of fashions (resembling Llama 405B) necessitates intricate multi-GPU orchestration as a result of they can’t be rendered on a single GPU.

Meet a cool start-up Felafax. Beginning with 8 TPU cores and going as much as 2048 cores, Felafax’s new cloud layer makes constructing AI coaching clusters easy. That can assist you get going quick, it provide pre-made templates for PyTorch XLA and JAX which can be straightforward to arrange. Simplified LLaMa Superb-tuning—use pre-built notebooks to leap proper into fine-tuning LLaMa 3.1 fashions (8B, 70B, and 405B). Felafax has taken care of the advanced multi-TPU orchestration.

A competing stack to NVIDIA’s CUDA, Felafax’s open-source AI platform is about to debut within the subsequent weeks. It’s based mostly on JAX and OpenXLA. They supply 30% cheaper efficiency than NVIDIA whereas supporting AI coaching on a variety of non-NVIDIA {hardware}, together with Google TPU, AWS Trainium, AMD, and Intel GPU.

Key Options

  • Giant coaching cluster with one click on: shortly spin up 8 to 1024 TPUs or non-Nvidia GPU clusters. Irrespective of the dimensions of the cluster, the framework effortlessly handles the coaching orchestration.
  • The bespoke coaching platform, constructed on a non-cuda XLA structure, affords unequalled efficiency at a decrease price. At 30% much less expense, you obtain the identical degree of efficiency as H100.
  • Personalize your coaching run by dropping it into your Jupyter pocket book on the contact of a button: full command, no room for error.
  • Felafax deal with all of the grunt work, together with optimizing mannequin partitioning for Llama 3.1 405B, coping with distributed checkpointing, and orchestrating coaching on a number of controllers. Redirect your consideration from infrastructure to innovation.
  • Customary templates: You could have two choices: Pytorch XLA and JAX. Use pre-configured environments with all of the required dependencies put in and get going instantly.
  • Llama 3.1’s JAX implementation: Coaching occasions are lowered by 25%, and GPU utilization is elevated by 20% utilizing JAX. Get essentially the most out of the costly computing you’ve invested in.
See also  Anthropic quietly expands access to Claude 'Private Alpha' at open-source event in San Francisco

In Conclusion

Felafax is developing an open-source AI platform to be used with next-gen AI expertise, which can lower the price of machine studying coaching by 30%. The group strives to make high-performance AI computing accessible to extra folks with its open-source platform and emphasis on GPUs that NVIDIA doesn’t make. There’s nonetheless a protracted method to go, however Felafax’s work might revolutionize synthetic intelligence by chopping prices, growing accessibility, and inspiring creativity. 


Source link

You may also like

logo

Welcome to our weekly AI News site, where we bring you the latest updates on artificial intelligence and its never-ending quest to take over the world! Yes, you heard it right – we’re not here to sugarcoat anything. Our tagline says it all: “because robots are taking over the world.”

Subscribe

Subscribe my Newsletter for new blog posts, tips & new photos. Let's stay updated!

© 2023 – All Right Reserved.