<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:cc="http://cyber.law.harvard.edu/rss/creativeCommonsRssModule.html">
    <channel>
        <title><![CDATA[Stories by HPC-AI Tech on Medium]]></title>
        <description><![CDATA[Stories by HPC-AI Tech on Medium]]></description>
        <link>https://medium.com/@hpcaitech?source=rss-ec86827abf4------2</link>
        <image>
            <url>https://cdn-images-1.medium.com/fit/c/150/150/1*aY_qlaO5o1gGk5B9-5LuFg.png</url>
            <title>Stories by HPC-AI Tech on Medium</title>
            <link>https://medium.com/@hpcaitech?source=rss-ec86827abf4------2</link>
        </image>
        <generator>Medium</generator>
        <lastBuildDate>Tue, 07 Apr 2026 15:00:49 GMT</lastBuildDate>
        <atom:link href="https://medium.com/@hpcaitech/feed" rel="self" type="application/rss+xml"/>
        <webMaster><![CDATA[yourfriends@medium.com]]></webMaster>
        <atom:link href="http://medium.superfeedr.com" rel="hub"/>
        <item>
            <title><![CDATA[HPC-AI Tech is Joining NYU’s Endless Frontier Labs Program, Which Has an Under 7% Global…]]></title>
            <link>https://medium.com/@hpcaitech/hpc-ai-tech-is-joining-nyus-endless-frontier-labs-program-which-has-an-under-7-global-927a2dd663a4?source=rss-ec86827abf4------2</link>
            <guid isPermaLink="false">https://medium.com/p/927a2dd663a4</guid>
            <category><![CDATA[high-performace-computing]]></category>
            <category><![CDATA[startup]]></category>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[ai]]></category>
            <dc:creator><![CDATA[HPC-AI Tech]]></dc:creator>
            <pubDate>Wed, 05 Oct 2022 09:47:22 GMT</pubDate>
            <atom:updated>2022-10-05T09:47:22.401Z</atom:updated>
            <content:encoded><![CDATA[<h3>HPC-AI Tech is Joining NYU’s Endless Frontier Labs Program, Which Has an Under 7% Global Acceptance Rate</h3><p>9 Sept, 2022 — <a href="https://hpcaitech.com/">HPC-AI Tech</a> was chosen from 1,121 applicants to join the <a href="https://endlessfrontierlabs.com/">Endless Frontier Labs (EFL)</a> 2022–2023 cohort Digital Tech track after a rigorous selection process by the EFL Team. NYU’s EFL program provides an opportunity for science and technology startups in their early stages, to grow in partnership with the New York University Stern School of Business. This year, over 1,100 startups from 66 countries world-wide, and from 43 states in the U.S. participated in the competition. HPC-AI Tech was among only 78 startups which were selected as finalists for the program.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*nHtf-_C8JmPJiyxWC2y8Hg.png" /></figure><h4><strong>More About EFL</strong></h4><p>Endless Frontier Labs (EFL) is a nine-month program at NYU’s Stern School of Business which provides early-stage science and technology startups with three tracks. These tracks include Deep Tech, Life Science, and Digital Tech. Deep Tech is suited for startups who develop technologies in the fields of physical and material science. Life Science is for companies with breakthroughs in biology, medicine and healthcare. Lastly, Digital Tech focuses on startups in software, data analytics, AI and ML.</p><p>Over the past few years, the EFL Team has utilized its extraordinary expertise to help startups transform their ideas into strong and successful businesses through the program. One example of innovative startups which have partnered with EFL include Immunai, which created an atlas of the human immune system in 2018 and recently raised about $295 million. Another example is Jetpack Aviation, a powerful company that specializes in turbine powered micro VTOLs; their 150mph Speeder flying motorcycles will hit the markets in 2023 for $380,000.</p><blockquote>“Our mission is to bridge the gap between science and markets,” said NYU Stern Professor <a href="https://www.stern.nyu.edu/faculty/bio/deepak-hegde">Deepak Hegde</a>, Founding Director of EFL. “We believe business strategy, validation by scientific peers, and connections to investors are critical for the successful commercialization of scientific breakthroughs. EFL provides these inputs to help transform founders’ ideas into commercial and societal impact. With the opening of our third track focused on software and data oriented startups, we look forward to deepening our impact on the information economy.”</blockquote><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*PFCeKTHq_Fp33UqQ7rpMAQ.png" /></figure><p>Companies must receive critical and extensive evaluations from the EFL Team in order to be apart of the <a href="https://endlessfrontierlabs.com/">EFL</a> program. The EFL team’s evaluation process included interviews with over 175 start-up teams from over 60 countries, technical analysis of the companies by their experts, and consideration for how well the company suited the EFL program.</p><p>Despite such stringent requirements, what made HPC-AI Tech stand out amongst 1,121 applicants?</p><p><a href="https://hpcaitech.com/">HPC-AI Tech</a> is a global company focused on High Performance Computing and Artificial Intelligence. The company has offices in Delaware (USA) and Singapore, building a highly skilled team comprised of members from top universities (e.g., the University of California, Berkeley, Stanford University, etc.) and from leading companies around the world (e.g., Google, Microsoft, Nvidia, IBM, Intel, etc.).</p><p>HPC-AI Tech strives to increase AI productivity, aiming to build an outstanding distributed AI development and deployment platform that enables supercomputers and cloud platforms to partake in AI at a much lower cost. Thanks to years of academic achievement and industrial accumulation, HPC-AI Tech has developed an efficient large AI model training and inference system — The <a href="https://github.com/hpcaitech/ColossalAI">Colossal-AI</a> — which offers efficient AI deployment services with high integration, automatization, and intellectualization. Recently, HPC-AI Tech completed the seed and angel round fundraising of 6 million dollars, led by BlueRun Ventures in the angel round, and co-led by Sinovation Ventures and ZhenFund in the seed round. HPC-AI Tech will continue its path of rapid growth and plans to continue expanding globally.</p><blockquote>“The team at Endless Frontier Labs is pleased to invite HPC-AI Technology Inc. to join our 2022–2023 Digital Tech track. Congratulations! We are thrilled for the opportunity to work with you and transform your breakthrough science into a highgrowth business.”</blockquote><blockquote>——Deepak Hegde, Founding Director of EFL</blockquote><blockquote>“As a startup focused on high-performance computing and Artificial Intelligence, we’re elated to join NYU’s <a href="https://endlessfrontierlabs.com/">Endless Frontier Labs</a>. We will passionately work with the professionals from the EFL Team to equip our innovations and form a successful business. This collaboration is another step towards achieving our goal of being a globally renowned AI company, and we hope we can empower AI businesses in the future. ”</blockquote><blockquote>— — Yang You, the Chairman of HPC-AI Tech</blockquote><h4><strong>Reference:</strong></h4><p>[1] <a href="https://endlessfrontierlabs.com/">https://endlessfrontierlabs.com/</a></p><p>[2] <a href="https://www.stern.nyu.edu/experience-stern/news-events/endless-frontier-labs-announces-2021-2022-application-launch-new-digital-tech-track-graduating">https://www.stern.nyu.edu/experience-stern/news-events/endless-frontier-labs-announces-2021-2022-application-launch-new-digital-tech-track-graduating</a></p><p>[3] <a href="https://www.swisspod.ch/stories/joining-nyus-endless-frontier-labs">https://www.swisspod.ch/stories/joining-nyus-endless-frontier-labs</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=927a2dd663a4" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[HPC-AI Tech Completes $6 Million Seed and Angel Round Fundraising, Led by BlueRun Ventures in the…]]></title>
            <link>https://medium.com/@hpcaitech/hpc-ai-tech-completes-6-million-seed-and-angel-round-fundraising-led-by-bluerun-ventures-in-the-892468cc2b02?source=rss-ec86827abf4------2</link>
            <guid isPermaLink="false">https://medium.com/p/892468cc2b02</guid>
            <category><![CDATA[angel-investors]]></category>
            <category><![CDATA[funding]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[high-performace-computing]]></category>
            <category><![CDATA[fundraising]]></category>
            <dc:creator><![CDATA[HPC-AI Tech]]></dc:creator>
            <pubDate>Tue, 06 Sep 2022 14:43:16 GMT</pubDate>
            <atom:updated>2022-09-06T14:43:16.421Z</atom:updated>
            <content:encoded><![CDATA[<h3>HPC-AI Tech Completes $6 Million Seed and Angel Round Fundraising, Led by BlueRun Ventures in the Angel Round</h3><p><strong>6 Sept, 2022 — </strong><a href="https://hpcaitech.com/">HPC-AI Tech</a>, the developer of open source deep learning system for big model era, <a href="https://github.com/hpcaitech/ColossalAI">Colossal-AI,</a> today announced the completion of seed and angel round fundraising of $6 million dollars, led by <a href="https://www.brv.com/">BlueRun Ventures</a> in the angel round, and co-led by <a href="https://www.sinovationventures.com/">Sinovation Ventures</a> and <a href="https://en.zhenfund.com/">ZhenFund</a> in the seed round. HPC-AI Tech completed two rounds of fundraising very quickly within one year. With this investment, HPC-AI Tech will continue its rapid growth by preparing for global expansion.</p><p>HPC-AI Tech is a global company focusing on High Performance Computing and Artificial Intelligence. The entities of HPC-AI Tech were registered in Delaware (USA) and Singapore. It has built a world-class team with core members from top universities (e.g., the University of California, Berkeley, Stanford University, etc.) and leading companies around the world (e.g., Google, Microsoft, Nvidia, IBM, Intel, etc.).</p><p>HPC-AI Tech strives to increase AI productivity and build a world-class distributed AI development and deployment platform that enables supercomputers and cloud platforms to serve AI at a much lower cost. Based on years of academic achievements and industrial accumulation, HPC-AI Tech has developed an efficient large AI model training and inference system,<a href="https://github.com/hpcaitech/ColossalAI"> Colossal-AI</a>, to deal with the pain points in the industry. HPC-AI Tech will be dedicated to the deep integration of software system design and hardware architecture, offering efficient AI deployment services in a manner of high integration, automatization, and intellectualization.</p><p>As cutting-edge AI models are getting larger, the budgets for training and deploying AI models are getting unaffordable both for enterprises and individuals. It can even lead to spending huge amounts of money on supercomputer clusters or hiring experts, it is difficult to apply large-scale AI models efficiently.</p><p>To this end, HPC-AI Tech has developed a user-friendly deep learning system, <a href="https://github.com/hpcaitech/ColossalAI">Colossal-AI</a>, that enables companies to maximize the efficiency of AI deployments while drastically reducing costs. Colossal-AI integrates many advanced technologies such as multi-dimensional tensor parallelism, sequence parallelism, heterogeneous memory management, large-scale optimization, adaptive task scheduling, etc.</p><p>Since being open source, Colossal-AI has reached NO.1 in trending projects on Github and Papers With Code several times, together with other projects that have as many as <strong>10K stars</strong>. Furthermore, Colossal-AI always keeps increasing the availability of AI solutions for industries and is already showing tremendous potential across a variety of fields including <strong>medicine, autonomous vehicles, cloud computing, retail, and chip production.</strong></p><p>“The global market size of AI is valued at hundreds of billions of dollars and it’s growing fast due to an increase in demand for improving efficiency and productivity of society. HPC-AI Tech’s solution can help enterprises and users effectively reduce the cost of big AI model training and inference, making large AI models more accessible, and thus benefiting wider society. We will continue to explore more effective industrial solutions of large AI models to democratize complicated large model techniques and empower AI enabled businesses. ”</p><p>— — Yang You, the Chairman of HPC-AI Tech</p><p>“With the rapid evolution and development of AI models, AI is capable of effectively solving various kinds of problems in the industry. However, enterprises face high costs in the training, maintenance, and inference of AI models. To enable customers to effectively integrate extensive AI capabilities into the enterprise lifecycle, HPC-AI Tech has developed an open source efficient deep learning system, that allows users to significantly improve model training and inference efficiency. As an early investor of HPC-AI Tech, BlueRun Ventures is glad to see that HPC-AI Tech has started to develop at a high speed and has been recognized by more and more users in terms of its product and global open source ecology. HPC-AI Tech is believed to grow into a globally renowned AI company and empower various industries with AI in the near future.”</p><p>— — Jimmy Shi, Venture Partnership at BlueRun Ventures China</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*NFKuJL4GbUllcXAm" /><figcaption><em>Yang YOU, the Chairman of HPC-AI Tech (left) and Jimmy Shi, Venture Partnership at BlueRun Ventures China (right)</em></figcaption></figure><p><strong>About HPC-AI Tech</strong></p><p>HPC-AI Tech is a global company focusing on High Performance Computing and Artificial Intelligence. The company was founded by Dr. Yang You, who is the Presidential Young Professor at the National University of Singapore and received his Ph.D. in Computer Science from UC Berkeley.</p><p>HPC-AI Tech has developed an efficient large AI model training and inference system, Colossal-AI, that integrates many advanced technologies such as multi-dimensional tensor parallelism, sequence parallelism, heterogeneous memory management, large-scale optimization, adaptive task scheduling, etc. By using Colossal-AI, we could help users to efficiently and quickly deploy large AI model training and inference, reducing large AI model training budgets and scaling down the labor cost of learning and deployment.</p><p><strong>Media Contact:</strong></p><p>HPC-AI Technology</p><p>Yang You</p><p>contact@hpcaitech.com</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=892468cc2b02" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Meet Colossal-AI Team at SC22 and Other 3 Renowned International Conferences]]></title>
            <link>https://medium.com/@hpcaitech/meet-colossal-ai-team-at-sc22-and-other-3-renowned-international-conferences-5b4e0b080cf?source=rss-ec86827abf4------2</link>
            <guid isPermaLink="false">https://medium.com/p/5b4e0b080cf</guid>
            <category><![CDATA[high-performace-computing]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[open-source]]></category>
            <category><![CDATA[computer-science]]></category>
            <dc:creator><![CDATA[HPC-AI Tech]]></dc:creator>
            <pubDate>Tue, 30 Aug 2022 14:41:21 GMT</pubDate>
            <atom:updated>2022-08-30T14:41:21.346Z</atom:updated>
            <content:encoded><![CDATA[<p>Recently, Colossal-AI Team, which developed a unified deep learning system for the big model era, has been accepted and invited to deliver keynote speeches at a series of notable international conferences including <strong>SuperComputing 2022 (SC22),</strong> <strong>Open Data Science Conference (ODSC), World Artificial Intelligence Conference (WAIC), and AWS Summit. </strong>In the event, Colossal-AI Team is going to share many up-to-date and amazing things and technologies of High Performance Computing (HPC) and Artificial Intelligence (AI) that will change the world. Follow us and stay tuned!</p><p><strong>Colossal-AI Open Source Code</strong>: <a href="https://github.com/hpcaitech/ColossalAI">https://github.com/hpcaitech/ColossalAI</a></p><h3>SuperComputing 2022 (SC22)</h3><p><strong>Time: Monday, 14 November 2022 1:30pm — 5pm CST</strong></p><p><strong>Link:</strong> <a href="https://sc22.supercomputing.org/presentation/?id=tut129&amp;sess=sess211">https://sc22.supercomputing.org/presentation/?id=tut129&amp;sess=sess211</a></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*X_0QC5CWd32DdVh2" /></figure><p><strong>SC</strong> (formerly <strong>Supercomputing</strong>), the <strong>International Conference for High Performance Computing</strong>, <strong>Networking</strong>, <strong>Storage and Analysis</strong>, is the annual conference established in 1988 by the <a href="https://en.wikipedia.org/wiki/Association_for_Computing_Machinery">Association for Computing Machinery</a> and the <a href="https://en.wikipedia.org/wiki/IEEE_Computer_Society">IEEE Computer Society</a>. SC brings together the world’s top research institutions and companies in the computer industry to share about the cutting-edge developments and innovations in HPC, networking, storage and analysis that will unlock new solutions and change our world.</p><h3>Open Data Science Conference (ODSC)</h3><p><strong>Time: 1 November — 3 November, 2022</strong></p><p><strong>Link:</strong> <a href="https://odsc.com/california/">https://odsc.com/california/</a></p><p>Open Data Science Conference (ODSC) hosts one of the largest gatherings of professional data scientists, with major conferences in the USA, Europe, and Asia. ODSC brings about the leading practitioners, innovation experts, and business professionals that drive artificial intelligence across a range of industries. During ODSC, attendees are able to learn more about insightful ideas and innovative AI techniques that will gain further momentum in different industries.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*oTbikzcdiZ6nh6pl" /></figure><h3>AWS Summit</h3><p><strong>Time: 22 September — 23 September, 2022</strong></p><p><strong>Link: </strong><a href="https://summit.awsevents.cn/2022/">https://summit.awsevents.cn/2022/</a></p><p>AWS Summit is the largest annual technology event of Amazon Cloud Technology in China. As a global cloud computing trendsetter event, AWS Summit gathers many leading technology practitioners and business professionals around the world to share their insightful views. At the summit, they will discuss how Amazon Cloud Technologies continues to explore the possibilities of cloud technology and empower different industries for more potential.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*c9eV0WZDnfmzWRyY" /></figure><h3>World Artificial Intelligence Conference (WAIC)</h3><p><strong>Time: 3 September, 2022</strong></p><p><strong>Link:</strong> <a href="https://www.worldaic.com.cn/">https://www.worldaic.com.cn/</a></p><p>World Artificial Intelligence Conference (WAIC) has become the most influential professional and international high-end exchange and cooperation platform that connects the development of artificial intelligence. Over the past three years, the event has invited nearly 300 high-profile speakers, including Turing Award winners, executives from leading multinational conglomerates, and founders of unicorn AI companies. During WAIC 2021, a total of more than 1,000 speakers shared their views and nearly 300 media reported the event.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*GM7XqmPbzFWW4csc" /></figure><h3>About us</h3><p><strong>HPC-AI Tech</strong></p><p>HPC-AI Tech is a global company focusing on High Performance Computing and Artificial Intelligence. The company was founded by Dr. Yang You, who is the Presidential Young Professor at the National University of Singapore and received his Ph.D. in Computer Science from UC Berkeley.</p><p>HPC-AI Tech has developed an efficient large AI model training and inference system, Colossal-AI, that integrates many advanced technologies such as multi-dimensional tensor parallelism, sequence parallelism, heterogeneous memory management, large-scale optimization, adaptive task scheduling, etc. By using Colossal-AI, we could help users to efficiently and quickly deploy large AI model training and inference, reducing large AI model training budgets and scaling down the labor cost of learning and deployment.</p><p><strong>Prof. Yang You</strong></p><p>Prof. Yang You is a Presidential Young Professor at the National University of Singapore. He received his Ph.D. in Computer Science from UC Berkeley. The focus of his current research is scaling up deep neural networks training on distributed systems or supercomputers. In 2017, his team broke the world record of ImageNet training speed, which was covered by the technology media like NSF, ScienceDaily, Science NewsLine, and i-programmer. In 2019, his team broke the world record of BERT training speed. The BERT training techniques have been used by many tech giants like Google, Microsoft, and NVIDIA. Yang You’s LARS and LAMB optimizers are available in industry benchmark MLPerf. He is a winner of IPDPS 2015 Best Paper Award (0.8%), ICPP 2018 Best Paper Award (0.3%) and ACM/IEEE George Michael HPC Fellowship. Yang You is a Siebel Scholar and a winner of Lotfi A. Zadeh Prize. Yang You was nominated by UC Berkeley for ACM Doctoral Dissertation Award (2 out of 81 Berkeley EECS PhD students graduated in 2020). He also made Forbes 30 Under 30 Asia list (2021) for young leaders and IEEE-CS TCHPC early career award.</p><p><strong>Prof. James Demmel</strong></p><p>Prof. James Demmel is the Dr. Richard Carl Dehmel Distinguished Professor of Computer Science and Mathematics at the University of California at Berkeley, and former Chair of the EECS Dept. He also serves as Chief Strategy Officer for the start-up HPC-AI Tech, whose goal is to make large-scale machine learning much more efficient, with little programming effort required by users. Demmel’s research is in high performance computing, numerical linear algebra, and communication avoiding algorithms. He is known for his work on the widely used LAPACK and ScaLAPACK linear algebra libraries. He is a member of the National Academy of Sciences, National Academy of Engineering, and American Academy of Arts and Sciences; a Fellow of the AAAS, ACM, AMS, IEEE and SIAM; and winner of the IPDPS Charles Babbage Award, IEEE Computer Society Sidney Fernbach Award, the ACM Paris Kanellakis Award, the J. H. Wilkinson Prize in Numerical Analysis and Scientific Computing, and numerous best paper prizes.</p><p><strong>Reference</strong></p><p><a href="https://www.163.com/dy/article/HEDANKUM0552XXBZ.html">https://www.163.com/dy/article/HEDANKUM0552XXBZ.html</a></p><p><a href="https://www.worldaic.com.cn/forum">https://www.worldaic.com.cn/forum</a></p><p><a href="https://odsc.com/california/events-west/">https://odsc.com/california/events-west/</a></p><p><a href="https://sc22.supercomputing.org/">https://sc22.supercomputing.org/</a></p><p><a href="https://mp.weixin.qq.com/s/kdjXzsaPdmCl8gAeEZ9_1w">https://mp.weixin.qq.com/s/kdjXzsaPdmCl8gAeEZ9_1w</a></p><p><a href="https://baike.baidu.com/item/%E5%85%A8%E7%90%83%E8%B6%85%E7%BA%A7%E8%AE%A1%E7%AE%97%E5%A4%A7%E4%BC%9A/16765988">https://baike.baidu.com/item/%E5%85%A8%E7%90%83%E8%B6%85%E7%BA%A7%E8%AE%A1%E7%AE%97%E5%A4%A7%E4%BC%9A/16765988</a></p><p><a href="https://www.baike.com/wikiid/7900314974072756281?from=wiki_content&amp;prd=innerlink&amp;view_id=1s5vstdve7b400">https://www.baike.com/wikiid/7900314974072756281?from=wiki_content&amp;prd=innerlink&amp;view_id=1s5vstdve7b400</a></p><p><a href="https://summit.awsevents.cn/2022/">https://summit.awsevents.cn/2022/</a></p><p><a href="https://en.wikipedia.org/wiki/ACM/IEEE_Supercomputing_Conference">https://en.wikipedia.org/wiki/ACM/IEEE_Supercomputing_Conference</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=5b4e0b080cf" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Accelerating Structure Prediction of Protein Monomers and Multimer by 11 Times!]]></title>
            <link>https://medium.com/@hpcaitech/accelerating-structure-prediction-of-protein-monomers-and-multimer-by-11-times-769715dcb5b5?source=rss-ec86827abf4------2</link>
            <guid isPermaLink="false">https://medium.com/p/769715dcb5b5</guid>
            <dc:creator><![CDATA[HPC-AI Tech]]></dc:creator>
            <pubDate>Tue, 23 Aug 2022 10:03:31 GMT</pubDate>
            <atom:updated>2022-08-23T10:03:31.577Z</atom:updated>
            <content:encoded><![CDATA[<h3><strong>Accelerating Structure Prediction of Protein Monomers and Multimer by 11 Times! An Open Source Solution from Colossal-AI and BioMap</strong></h3><p>The latest solution from the Colossal-AI team (<a href="https://github.com/hpcaitech/ColossalAI">https://github.com/hpcaitech/ColossalAI</a>) and BioMap for protein monomer and multimer structure prediction, xTrimo Multimer, has recently become open-source to the public. This new solution can predict <strong>both monomer and multimer structure</strong> simultaneously accelerating the process by up to <strong>11 times</strong>!</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/852/1*ziOxbOQprGaxTEfSe2uiGg.png" /></figure><p>The hero behind is Colossal-AI, which is a powerful deep learning system that aims to make large AI model training easy and accessible in the community and industry. By integrating large model training techniques and optimizations provided by Colossal-AI, we can significantly reduce the time and cost of both protein monomer and multimer structure prediction during model training and inference. As an important practice of the Colossal-AI system in the pharmaceutical industry, xTrimo Multimer can greatly increase the pace of the model design and development for protein structure prediction, facilitating breakthroughs regarding large AI model applications in healthcare and bioinformatics.</p><p>Learn more about our powerful solutions here: <a href="https://github.com/hpcaitech/ColossalAI/#xTrimoMultimer">https://github.com/hpcaitech/ColossalAI/#xTrimoMultimer</a></p><p>Colossal-AI is a user-friendly deep learning system that enables companies to maximize the efficiency of AI deployments while drastically reducing costs. Since open source to the public, Colossal-AI has reached №1 in trending projects on Github and Papers With Code several times, together with other projects that have as many as 10K stars. Furthermore, Colossal-AI always keeps increasing the availability of AI solutions for industries and is already showing tremendous potential across a variety of fields including <strong>medicine, autonomous vehicles, cloud computing, retail, and chip production. </strong>The most recent application is the partnership of Colossal-AI with BioMap to propose the latest cost-effective solution in protein monomer and multimer structure prediction. This application is able to help healthcare providers and pharmaceutical companies in diagnosis and stimulate novel drug research and discovery.</p><p>Protein structure prediction is one of the most important topics in structural biology and supplements the understanding of gene translation and protein function. Unfortunately, the multi-level structure and sophisticated protein interactions make it challenging to predict the 3D structure accurately.</p><p>In recent years, the success of deep neural networks has transformed various practices. Since DeepMind’s release of AlphaFold (which accurately predicts protein structure based on amino acid sequences), the field of biology has witnessed a boom in utilization of AI for protein structure prediction.</p><p>Specifically, AlphaFold can generate end-to-end 3D structure predictions of protein monomers directly from amino acid sequences. The use of AlphaFold also exceeds the realm of monomers. Since the majority of proteins function as multimers, DeepMind’s AlphaFold-Multimer model is recently released to be able to predict the structure of multimers.</p><p>To boost the development of AlphaFold, the Colossal-AI team has already released FastFold, an open-source and optimized implementation of AlphaFold in the past few months. The Colossal-AI team managed to successfully minimize AlphaFold’s training time from 11 days to only 67 hours abd accelerate inference time by up to ~11.6 times. The Colossal-AI team continues their efforts to democratize large-scale AI model applications in the pharmaceutical field.</p><p>Interactions between proteins are critical to their biological functions. To address difficulties related to protein monomer and multimer structure prediction, the Colossal-AI Team proposed the industry’s latest solution, the xTrimo Multimer. xTrimo Multimer is able to better reflect protein interactions, thus enhancing the potential target analysis, protein structure prediction/ simulation, as well as high-precision antibody designs in drug discovery and development.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/852/1*7YYfBXpv3TzjIcCm2Zol1A.jpeg" /></figure><p>The unaffordable economic and time costs from AlphaFold’s inference has led to challenges in its research and development, particularly when facing long sequence inferences with rising computational complexity and memory consumption. Based upon computational features in the AlphaFold-Multimer model, the Colossal-AI team introduced CUDA optimization and Kernel Fusion techniques for the xTrimo Multimer achieving remarkable inference performances. Compared to AlphaFold2 and OpenFold (from Columbia University), the xTrimo Multimer has a significantly improved inference performance on a single GPU by 1.58–2.14 times and 1.14–2.23 times, respectively.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/852/1*QvwRRHq0oZoBL509toFYlQ.png" /></figure><p>Additionally, the xTrimo Multimer model supports distributed inferences for lengthy sequences. After introducing Dynamic Axial Parallelism, the xTrimo Multimer was able to efficiently distribute computation and partial GPU memory across a variety of devices, thereby solving computational and memory challenges that long sequences face. xTrimo Multimer achieves a 8.47x and 11.15x speedup compared to OpenFold and AlphaFold 2 on multiple GPUs with sequence lengths ranging anywhere from 2–3K. xTrimo Multimer also shows a 4.45x acceleration compared to Uni-Fold 2.0. Furthermore, xTrimo Multimer can support inferences with sequences reaching up to 4K, whereas OpenFold and AlphaFold 2 are restricted from such lengths due to GPU memory. With xTrimo Multimer, scientists can run a 4K length sequence inference in about 20 minutes.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/852/1*q89vx01y5gsxUIxSKpgATw.png" /></figure><p>“The collaboration with HPC-AI Tech brings together the cutting-edge technology in large AI model training from Colossal-AI team and the biocomputing domain expertise from BioMap. The release of xTrimo Multimer is an important step towards integrating the advantages of large AI model training and inference into the construction of BioMap’s xTrimo multimodal system. ” — Le Song, Chief AI Scientist of BioMap</p><p>“Our latest solution of protein Monomer and Multimer structure prediction is an important progress of Colossal-AI to solve the industrial problems in the real world. In the future, we will continue to cooperate with BioMap more deeply in biocomputing large models, to stimulate the application and implementation of deep learning in innovative drug development. ” — Yang You, the Chairman of HPC-AI Tech</p><p>The accelerated implementation, xTrimo Multimer, will serve as one of the important products, alongside other impressive industrial solutions built upon Colossal-AI, to facilitate the large-scale AI modeling for global companies. The Colossal-AI team will continue to explore all the emerging possibilities of AI model training in various fields, endeavor to tackle issues in the modern industry, and empower the future of the global AI market.</p><p><strong>Portal</strong></p><p>Open Source xTrimoMultimer: <a href="https://github.com/hpcaitech/ColossalAI/#xTrimoMultimer">https://github.com/hpcaitech/ColossalAI/#xTrimoMultimer</a></p><p>Open Source Colossal-AI: <a href="https://github.com/hpcaitech/ColossalAI">https://github.com/hpcaitech/ColossalAI</a></p><p><strong>About BioMap</strong></p><p>BioMap is a team of world-renowned scientists who have extensive expertise in disease biology, bioinformatics, machine learning/deep learning, and antibody engineering. BioMap was co-founded by Baidu’s Founder/CEO Robin Li and the former CEO of Baidu Ventures, Wei Liu. They are committed to bringing first-in-class medicine for unmet medical needs in the areas of immune-oncology, autoimmune diseases, fibrosis and aging-related diseases.</p><p><strong>About HPC-AI Tech</strong></p><p>HPC-AI Tech is a global company which aims to help users improve the efficiency of training and deploying large AI models. The company was founded by Dr. Yang You, who received his Ph.D. in Computer Science from UC Berkeley and is currently the Presidential Young Professor at the National University of Singapore. HPC-AI Tech has developed an efficient large AI model training and inference system, Colossal-AI, that integrates advanced technologies which help users efficiently deploy large AI model training and inference at a low cost.</p><p><strong>Reference:</strong></p><p><a href="https://www.biomap.com/en/team">https://www.biomap.com/en/team</a></p><p><a href="https://www.technologyreview.com/2021/07/22/1029973/deepmind-alphafold-protein-folding-biology-disease-drugs-proteome/">https://www.technologyreview.com/2021/07/22/1029973/deepmind-alphafold-protein-folding-biology-disease-drugs-proteome/</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=769715dcb5b5" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Job Opening: Sales Director]]></title>
            <link>https://medium.com/@hpcaitech/job-opening-sales-director-1f0f807f0f1f?source=rss-ec86827abf4------2</link>
            <guid isPermaLink="false">https://medium.com/p/1f0f807f0f1f</guid>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[hiring]]></category>
            <category><![CDATA[marketing]]></category>
            <category><![CDATA[sales]]></category>
            <category><![CDATA[deep-learning]]></category>
            <dc:creator><![CDATA[HPC-AI Tech]]></dc:creator>
            <pubDate>Fri, 29 Jul 2022 13:47:24 GMT</pubDate>
            <atom:updated>2022-09-02T03:21:39.347Z</atom:updated>
            <content:encoded><![CDATA[<p><a href="https://www.hpcaitech.com/">HPC-AI Tech</a> is a dynamic startup that is blazing the way in applying high-performance computing techniques to artificial intelligence. We are pioneers in writing performant, extensible and easy-to-use AI systems that run on the cloud.</p><p>Now, HPC-AI Tech is looking for an experienced sales director. He/She should manage and oversee the sales operations in our company. His/Her main duties include keeping good relationships with key clients and customers, designing and articulating plans to meet sales targets, evaluating costs to determine our products’ price, keeping abreast of AI era trends, and training sales managers.</p><p>HPC-AI Tech provides a high basic salary (<strong>80k — 160k USD</strong> per year) + a high sales commission (<strong>15% of each order</strong>) or a bonus (<strong>80k — 160k USD </strong>per year). We hope you work in the U.S.</p><p>Come and join us!</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*15nzog5gvLFgQ106" /></figure><p><strong>Roles and Responsibilities</strong></p><ol><li>Developing the company’s marketing strategy and helping the company’s products to be applied in the industry.</li><li>Responsible for the sales of the company’s products to customers, developing and executing strategic plans to achieve sales targets.</li><li>Building and maintaining long-term and strong relationships with customers while partnering with them to better understand their business objectives and needs.</li><li>Understanding industry-specific trends and landscapes, matching customer needs with solutions, and defining a clear path to client satisfaction and revenue growth for the company.</li><li>Increasing brand awareness and market share of the company’s products.</li></ol><p><strong>Qualifications</strong></p><ol><li>Bachelor’s degree in business, marketing, communications, or related fields.</li><li>At least 3-year B2B sales experience in AI, Cloud Computing, or related fields and a basic understanding of AI. Familiarity with industry-specific trends and landscapes of High-Performance Computing and AI will be advantageous.</li><li>Proven performance in industry sales and rich accumulation of convertible industry customer resources in AI, Internet, medical, automotive, and other related fields.</li><li>Excellent listening, negotiation, presentation, and communication skills</li><li>Experienced, energetic, and able to demonstrate the ability to sell.</li></ol><p><strong>Join Us</strong>：contact@hpcaitech.com</p><p>Image Source: <a href="https://pixabay.com/zh/photos/ladder-success-success-ladder-2713346/">https://pixabay.com/zh/photos/ladder-success-success-ladder-2713346/</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=1f0f807f0f1f" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[HPC-AI Tech Joins NVIDIA Inception]]></title>
            <link>https://medium.com/@hpcaitech/hpc-ai-tech-joins-nvidia-inception-798e057fe76f?source=rss-ec86827abf4------2</link>
            <guid isPermaLink="false">https://medium.com/p/798e057fe76f</guid>
            <category><![CDATA[high-performace-computing]]></category>
            <category><![CDATA[open-source]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[deep-learning]]></category>
            <dc:creator><![CDATA[HPC-AI Tech]]></dc:creator>
            <pubDate>Tue, 28 Jun 2022 13:00:15 GMT</pubDate>
            <atom:updated>2022-06-28T13:00:15.771Z</atom:updated>
            <content:encoded><![CDATA[<p><strong>June, 2022 — </strong><a href="https://hpcaitech.com/">HPC-AI Tech</a> today announced it has joined <a href="https://nvda.ws/2BvtUc9">NVIDIA Inception</a>, a program designed to nurture startups revolutionizing industries with technology advancements.</p><p>HPC-AI Tech is focused on increasing AI productivity and building a world-class distributed AI development and deployment platform that enables supercomputers and cloud platforms to serve AI at a much lower cost. Based on years of academic achievements and industrial accumulation, HPC-AI Tech has developed an efficient large AI model training and inference system, <a href="https://github.com/hpcaitech/ColossalAI">Colossal-AI</a>, to deal with the pain points in the industry. HPC-AI Tech will be dedicated to the deep integration of software system design and hardware architecture, offering efficient AI deployment services in a manner of high integration, automatization, and intellectualization.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*6PlwYnlc6Y1p1NOF" /></figure><p>NVIDIA Inception will allow HPC-AI Tech to evolve faster through access to cutting-edge technology and NVIDIA experts, connections with venture capitalists, and co-marketing support to heighten HPC-AI Tech’s visibility. NVIDIA will work closely with HPC-AI Tech to provide the best technical tools, latest resources, and opportunities to connect with investors. The program will also offer HPC-AI Tech the opportunity to collaborate with industry-leading experts and other AI-driven organizations.</p><blockquote><em>“We feel very happy to join NVIDIA Inception. It is such an impressive program that helps us to drive our business forward through go-to-market support, training, technical assistance, etc. We will work with NVIDIA to explore more effective applications of GPUs to democratize complicated large model techniques in the AI community.” — Yang You, the Chairman of HPC-AI Tech</em></blockquote><p>NVIDIA Inception helps startups during critical stages of product development, prototyping and deployment. Every NVIDIA Inception member gets a custom set of ongoing benefits, such as <a href="https://www.nvidia.com/en-us/deep-learning-ai/education/?ncid=em-ded-n1-96486">NVIDIA Deep Learning Institute</a> credits, marketing support, and technology assistance, which provides startups with the fundamental tools to help them grow.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*NNwNje1sQqNMJ_Bf" /></figure><p><strong>About HPC-AI Tech</strong></p><p>HPC-AI Tech is a global company focusing on High Performance Computing and Artificial Intelligence. The company was founded by Dr. Yang You, who is the Presidential Young Professor at the National University of Singapore and received his Ph.D. in Computer Science from UC Berkeley.</p><p>HPC-AI Tech has developed an efficient large AI model training and inference system, Colossal-AI, that integrates many advanced technologies such as multi-dimensional tensor parallelism, sequence parallelism, heterogeneous memory management, large-scale optimization, adaptive task scheduling, etc. By using Colossal-AI, we could help users to efficiently and quickly deploy large AI model training and inference, reducing large AI model training budgets and scaling down the labor cost of learning and deployment.</p><p><strong>Media Contact:</strong></p><p>HPC-AI Technology</p><p>Yang You</p><p>contact@hpcaitech.com</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=798e057fe76f" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Surpassing NVIDIA FasterTransformer’s Inference Performance by 50%, Open Source Project Powers into…]]></title>
            <link>https://medium.com/@hpcaitech/surpassing-nvidia-fastertransformers-inference-performance-by-50-open-source-project-powers-into-6139c5bc7790?source=rss-ec86827abf4------2</link>
            <guid isPermaLink="false">https://medium.com/p/6139c5bc7790</guid>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[open-source]]></category>
            <category><![CDATA[deep-learning]]></category>
            <dc:creator><![CDATA[HPC-AI Tech]]></dc:creator>
            <pubDate>Mon, 30 May 2022 23:53:55 GMT</pubDate>
            <atom:updated>2022-05-30T23:53:55.320Z</atom:updated>
            <content:encoded><![CDATA[<h3>Surpassing NVIDIA FasterTransformer’s Inference Performance by 50%, Open Source Project Powers into the Future of Large Models Industrialization</h3><p>The development of artificial intelligence(AI) and high-performance computing(HPC) has dramatically revolutionized our lives and brought significant potential to our society. For example, Facebook has 1.82 billion daily active users and issues tens of trillions of inference queries per day, which requires essential changes in the development process and system design.</p><p>Traditional deep learning(DL) systems usually focus on the single-model single-machine inference setting. However, the exponential growth of DL models leads to an inability to execute large pre-trained models on a single machine. Specifically, super large NLP models like GPT-3 require more than one hundred GBs of memory for inference. One single GPU is not enough to hold such tasks, which makes using multiple distributed computing devices to inference collaboratively the future of large model inference.</p><p>The Colossal-AI team developed Energon-AI, as a subsystem, to provide inference service for super-scale DL models. Focusing on the specific pain points, Colossal-AI Team delved into multi-device inference situations, developed large scale inference system Energon-AI, with their idea of “High Performance, High Usability, High Versatility”.</p><p><strong>With few changes to existing projects, users could easily develop large models for inference, achieve superlinear speedups on parallel extensions. Compared to FastTransformer introduced by NVIDIA, Energon-AI managed to reach an improvement of 50% on parallelized inference speedups with large AI models.</strong></p><p>Besides, different from current inference solutions, Energon-AI does not require manual settings for communication or memory usage, neither does extra compiling. By using Energon-AI, inference becomes way easier than before.</p><p>Open source address: <a href="https://github.com/hpcaitech/ColossalAI">https://github.com/hpcaitech/ColossalAI</a></p><p><strong>Difficulties with Large AI Model Inference</strong></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/563/0*lLVjZWoqYhSu6a08" /><figcaption>Rapid growth of model parameters [https://arxiv.org/abs/2111.14247]</figcaption></figure><p>Recently, computing devices such as GPUs have been greatly upgraded focusing on their parallel computing capability, memory capacity, memory speed, etc. However, such performance improvements on a single device could never meet the requirements of large models whose parameter size grows exponentially. Current deep learning inference systems mainly focus on simple scenarios like multi-model single-machine and single-model single-machine, overlooking the challenges and opportunities of single-model multi-machine scenarios that are essential for large AI model inference. Thus, we introduced Energon-AI system to solve these severe issues.</p><h3>Large Scale Inference System: Energon- AI</h3><h4><strong>System Design</strong></h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/645/0*KDiIVhDUBOFB08j5" /><figcaption>Schematic diagram of Energon-AI</figcaption></figure><p>Focusing on the deployment of large AI models, we design a single-model multi-machine inference system, Energon-AI. Its <strong>system is comprised of three levels, namely runtime system (Runtime), distributed inference instance (Engine) and service system (Serving)</strong>:</p><ul><li><strong>Runtime:</strong> In the design of the runtime system, we find that with tremendous model scaling, the time taken by general matrix multiplications gradually increases. Conversely, the time taken by memory-intensive operators and Kernel Launch shows a decreasing pattern. The process is further migrated from memory-intensive to computing-intensive, and the effect of TensorRT and specified inference systems for memory-intensive operations is greatly reduced. Energon-AI Runtime relies on Colossal-AI to achieve tensor parallelism. Meanwhile, we also design a pipeline parallel packaging method for insufficient memory. In addition, we introduce a large number of special inference operators and methods. For example, when it comes to the variable length of input in NLP, we introduce operators such as transpose_padding_rebulid and transpose_padding_remove to efficiently process MLP layers in the Encoder and Decoder models.</li><li><strong>Engine:</strong> To make Engine have exactly the same behavior as single-device inference through encapsulation, we adapt a semi-centralized method and use RPC to call initialization or inference methods on each device in the main process, so that distributed inference can be centrally controlled. Meanwhile, each device maintains its own communication logic for Tensor Parallel and Pipeline Parallel. We design and maintain a distributed message queue in each process to ensure the consistency of multi-threaded call execution in multiple processes.</li><li><strong>Serving:</strong> Energon-AI introduces a dynamic batching mechanism. After the requests in the request queue are optimally packaged according to machine performance, Energon-AI selects the batch processing with the highest priority regarding the waiting time, batch size, batch expansion possibility (based on the sentence length after padding), etc., which enables the maximization of the GPU usage while avoiding the starvation problem and reducing the average request delay.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/968/0*jL7Pqc8YcygPvPKF" /><figcaption>Schematic diagram of batch management process</figcaption></figure><h3>Performance Testing</h3><h4>Superlinear Scaling of Parallel Inference</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*rG5r34uxCibG-LPF" /><figcaption><em>Tensor parallel scalability test results. Hardware: 8*A100 GPU 80GB. Since the memory of a single device cannot meet the inference requirements of GPT-3, above is the test result of GPT-3 with 12 layers, and the sentence length is 1/2 of Padding.</em></figcaption></figure><p>When the Batch Size is 32, Energon-AI parallel inference with 8 GPUs can achieve an <strong>8.5-time super-linear speedup compared to </strong>Pytorch direct inference with one GPU.</p><h4>Runtime Inference Performance Improved by 50%</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*KrGW7UIxNocG30BV" /><figcaption><em>Comparison of inference latency of tensor parallel runtime systems. Hardware: 8*A100 GPU 80GB. Sentence length is set to be 1/2 length of Padding. GPT-3–24-Layers for TP=2, GPT-3–48-Layers for TP=4.</em></figcaption></figure><p>We compare our Energon-AI with highly optimized FasterTransformer GPT-3 introduced by NVIDIA. Faster Transformer introduces its distributed inference feature in its 4.0 version, and currently supports distributed inference of the GPT-3 model. However, due to its highly coupled pure C++ code, its flexibility and usability are relatively low. In addition, regarding the characteristics of different lengths of input sentences for NLP inference, its distributed inference has no redundant computation elimination method.</p><p>For the GPT-3 model, the runtime system of Energon-AI performs slightly lower than FasterTransformer when the batch size is 1, while it could achieve <strong>more than 50% performance improvement when the batch size increases</strong>.</p><h4>30% Increase in Dynamic Batching Throughput</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*SpxiXs_-_bNXJV5Y" /><figcaption><em>Dynamic batching versus direct packaging batch throughput. Hardware: 8*A100 GPU 80GB. The model used in the test is GPT-3, the test sentence length is randomly generated with a maximum of 256, and the padding strategy conforms to the longest padding in the batch.</em></figcaption></figure><p>We simulate a real-world scenario where multiple users send a large number of vary-length inference requests at the same time, and compare the throughput of our dynamic batch method with the traditional FIFO (first-in-first-out) method. Since the dynamic batching algorithm alleviates the problem of massive redundant computation caused by direct padding, the throughput of dynamic batching is improved by 34.7%.</p><h3>High Usability</h3><pre>from gpt import gpt3</pre><pre>from gpt_batch_server import launch_engine</pre><pre># for engine</pre><pre>model_class = gpt3</pre><pre>model_type = &quot;gpt&quot;</pre><pre>host = &quot;127.0.0.1&quot;</pre><pre>port = 29401</pre><pre>half = True</pre><pre>backend = &quot;nccl&quot;</pre><pre># for parallel</pre><pre>tp_init_size = 4</pre><pre>pp_init_size = 2</pre><pre># for server</pre><pre>engine_server = launch_engine</pre><pre>server_host = &quot;127.0.0.1&quot;</pre><pre>server_port = 8020</pre><pre>rm_padding = False</pre><pre>energonai service init --config_file=gpt_config.py</pre><p>While pursuing high performance, Energon-AI hopes to maintain high flexibility and usability of the system. Users only need to customize [parallel model], [parallel parameters] and [service request logic] in the configuration file to start an inference task. Currently, we provide the most common GPT, BERT and ViT models as examples, and more detailed tutorials will be provided in the near future.</p><p>When building a new parallel model, Energon-AI uses Python, and the usage is similar to Pytorch. It has the concept of layers and its logic of initialization and execution is clear. Users do not need to consider memory management or parallel communication. The following codes show how to run a model with two Linear layers in parallel with Energon-AI.</p><pre>class MLP(nn.Module):</pre><pre>def __init__(self, dim, dtype, bias):</pre><pre>super().__init__()</pre><pre>self.dense_0 = Linear1D_Col(dim, dim, dtype=dtype, bias=bias, gather_output=False)</pre><pre>self.dense_1 = Linear1D_Row(dim, dim, dtype=dtype, bias=bias, parallel_input=True)</pre><pre>def forward(self, x):</pre><pre>x = self.dense_0(x)</pre><pre>x = self.dense_1(x)</pre><pre>return x</pre><p>In contrast, when using FasterTransformer to build a new parallel model, users are required to write in C++ and manually set underlying behavior organizations such as managing memory, defining communication, etc. Due to space limitations, the following code shows the part needed for memory management, and communication settings of a two Linear layer model running in parallel with FasterTransformer. Besides, users need to spend a lot of time and energy to debug if they want the code to execute correctly, and C++ code needs additional compilation works. These all pose serious challenges to the user’s parallel knowledge and programming ability.</p><pre>// Memory Allocation (only for a single paramerter).</pre><pre>T *d_inter_kernel = NULL</pre><pre>param_.ffn.intermediate_weight.kernel = d_inter_kernel;</pre><pre>device_malloc(&amp;d_inter_kernel, dim * dim);</pre><pre>// Two MLP Layers</pre><pre>cublasMM_cublasLtMM_wrapper(param_.cublaslt_handle, param_.cublas_handle, CUBLAS_OP_N, CUBLAS_OP_N, n, m, k, &amp;alpha, param_.ffn.intermediate_weight.kernel, AType_, n, attr_matmul_buf_, BType_, k, &amp;beta, (DataType_ *)inter_matmul_buf_, CType_, n, param_.stream, cublasAlgoMap_, sm_, cublas_workspace_);</pre><pre>add_bias_act_kernelLauncher&lt;DataType_&gt;(inter_matmul_buf_, param_.ffn.intermediate_weight.bias, m, n, ActivationType::GELU, param_.stream);</pre><pre>n = k;</pre><pre>cublasMM_cublasLtMM_wrapper(param_.cublaslt_handle, param_.cublas_handle, CUBLAS_OP_N, CUBLAS_OP_N, n, m, k, &amp;alpha, param_.ffn.output_weight.kernel, AType_, n, inter_matmul_buf_, BType_, k, &amp;beta, (DataType_ *)(param_.transformer_out), CType_, n, param_.stream, cublasAlgoMap_, sm_, cublas_workspace_);</pre><pre>add_bias_input_layernorm_kernelLauncher&lt;DataType_&gt;(param_.transformer_out,                                                             attr_matmul_buf_, param_.ffn.output_weight.bias, param_.ffn_layernorm.gamma, param_.ffn_layernorm.beta, m, n, param_.stream);</pre><pre>// Communication</pre><pre>if(t_parallel_param_.world_size &gt; 1)</pre><pre>{</pre><pre>all2all_gather(nccl_logits_buf_, nccl_logits_buf_, local_batch * n, t_parallel_param_, decoding_params.stream);</pre><pre>}</pre><h3>More Features</h3><p>The Energon-AI subsystem we released recently is beta version. Based on user feedback and plans on schedule, intensive iterative updates will be carried out gradually, and the official version will be provided to users as soon as possible to fully meet the different inference deployment needs. Your requirements and suggestions for Energon-AI are always welcomed here.</p><h3>Building a Large AI Model Ecosystem</h3><p>Living in an age of large AI models, in order to solve the pain points of the existing solutions such as limited parallel dimension, low efficiency, poor versatility, difficult deployment, and lack of maintenance, Colossal-AI uses technologies such as efficient multi-dimensional parallelism and heterogeneous parallelism to <strong>allow users to deploy large AI models efficiently and quickly with only a few modifications of their codes</strong>.</p><p>For example, for a super-large AI model such as GPT-3, compared to the NVIDIA solution, Colossal-AI <strong>only needs half the computing resources</strong>; if the same computing resources are used, the speed could be further increased by 11%, which could <strong>reduce the training cost of GPT-3 over a million dollars</strong>.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/977/0*UcHMdtU4AoDE92ce" /></figure><p>For AlphaFold, which is used for protein structure prediction, our team has introduced FastFold based on the Colossal-AI acceleration scheme. FastFold successfully surpassed other schemes proposed by Google and Columbia University, reducing the training time of AlphaFold from 11 days to 67 hours, and the total cost is lowered as well. Also, in long sequence inference, we achieved a speed improvement of 9.3 ~ 11.6 times.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/977/0*PBbvKa5h18PDz_o4" /></figure><p>Colossal-AI is compatible with low-end devices and can <strong>train GPT with up to 18 billion parameters on a PC with only one GPU</strong>; ordinary laptops can also train models with more than one billion parameters. Compared with current popular solutions, the parameter capacity can be increased by more than 10 times, which greatly reduces the threshold for downstream tasks and application deployment such as the fine-tuning and inference of AI large model.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/750/0*Q8SJt2o2LMUz2T5X" /></figure><p>Besides, Colossal-AI values open source community construction, providing English and Chinese tutorials, and supporting the latest cutting-edge applications such as PaLM and AlphaFold. Colossal-AI will roll out new and innovative features regularly as well. We always welcome suggestions and discussions from the community, and we would be more than willing to help you if you encounter any issues.</p><p>Recently, Colossal-AI reached <strong>N.01 on the top trending projects on Github</strong>, against a backdrop of many projects that have as many as 10K stars.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/825/0*YXhtxU5tvxY8rQ3i" /></figure><p>On the Papers With Code, a website which highlights trending Machine Learning research, Colossal-AI has also topped the trending list and attracted many researchers’ attention with its amazing features.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*Mha4NCBFC7nciEF9" /></figure><h3>Portal</h3><p>Project address: <a href="https://github.com/hpcaitech/ColossalAI">https://github.com/hpcaitech/ColossalAI</a></p><h3>Funding</h3><p>HPC-AI Tech raised 4.7 million USD from top VC firms in just 3 months after the company was founded. For more information, please email contact@hpcaitech.com.</p><p>Reference:</p><p><a href="https://arxiv.org/pdf/2111.14247.pdf">https://arxiv.org/pdf/2111.14247.pdf</a></p><p><a href="https://www.sciencedirect.com/science/article/abs/pii/S0950584920301373">https://www.sciencedirect.com/science/article/abs/pii/S0950584920301373</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=6139c5bc7790" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Implementing the Gargantuan Pathways with Colossal-AI, easy and efficient!]]></title>
            <link>https://medium.com/@hpcaitech/implementing-the-gargantuan-pathways-with-colossal-ai-easy-and-efficient-6362c1144e53?source=rss-ec86827abf4------2</link>
            <guid isPermaLink="false">https://medium.com/p/6362c1144e53</guid>
            <category><![CDATA[distributed-systems]]></category>
            <category><![CDATA[parallel-computing]]></category>
            <category><![CDATA[google]]></category>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <dc:creator><![CDATA[HPC-AI Tech]]></dc:creator>
            <pubDate>Wed, 18 May 2022 06:25:19 GMT</pubDate>
            <atom:updated>2022-05-18T06:25:19.475Z</atom:updated>
            <content:encoded><![CDATA[<p>Today’s AI models are great at specializing at one task. Whether that be object detection over a particular domain of labels (for example cats and dogs), or generating natural language (e.g. a book or an essay) based on a prompt. To ask a model to do both, however, and you get into very murky territory. That is until 2021 where Google’s Jeff Dean came up with what may be the future of deep learning: Pathways. Pathways is a multi-modal, sparse, deep learning architecture that can <em>generalize to millions of tasks. </em>It is an incremental step towards a more artificially general and intelligent machine.</p><p>Nevertheless, compared to the original Transformers architecture, Pathways Language Model(PaLM) has made a series of bold innovations. However, due to its inordinate complexity, it, once again, places the burden on programmers to implement it efficiently on modern hardware. Moreover, significant parts of PaLM are not open sourced, making it difficult to implement parts of it on available hardware, such as GPUs.</p><p>Fret not! The Colossal-AI team has implemented the model structure of PaLM and applied several state-of-the-art (sota) high performance computing (HPC) techniques to its implementation, in a parallelised manner, across GPUs, extracting the very last iota of performance.</p><p>Let’s talk about the Colossal-AI project, Pathways and our efficient implementation.</p><p>If you feel interested, do check out our GitHub repo: <a href="https://github.com/hpcaitech/PaLM-colossalai">https://github.com/hpcaitech/PaLM-colossalai</a></p><p><strong>About Colossal-AI</strong></p><p>Colossal-AI is a deep learning system that makes it easy to write performant parallel code for your large AI models. It is based on PyTorch and provides an entire ecosystem to scale up your training to the next level.</p><p>GitHub page: <a href="https://github.com/hpcaitech/ColossalAI">https://github.com/hpcaitech/ColossalAI</a></p><p>Colossal-AI’s claim to fame is that it supports a variety of distribution methods, including tensor parallelism, pipeline parallelism, zero-redundant data parallelism etc… We have provided, as examples, efficient implementations of BERT, GPT and ViT that support hybrid levels of parallelism (combining multiple distributed acceleration methods at once). Evolving with the machine learning community, we now attempt to support the PaLM model’s distributed training.</p><p><strong>About Pathways</strong></p><p>Pathways is the next stage in the evolution of deep learning architectures. It is a large multi-modal system that is capable of generalizing across as many as millions of tasks. Unfortunately, Pathways is built for Google’s needs. As such, it is built over extremely special hardware (TPUs) and over its own special network as well. These mean that, whilst efficient in the Google ecosystem, additional works need to be carried out to scale to <em>accessible</em> hardwares (GPUs) for the society. The model size of PaLM is quite huge with 540 billion parameters, which strongly affects the efficiency a naive data parallel method can grant. The question that begs to be answered is: can we use our common GPU clusters to implement a performant PaLM Model? Well, the engineers of HPC-AI Tech have proposed an answer with Colossal-AI.</p><p>But before we delve into the implementation details, let’s talk about the intricacies of what makes the PaLM model great.</p><p><strong>PaLM Model Explained</strong></p><p>Compared to the regular Transformer architecture, PaLM innovates primarily in following ways:</p><ol><li>It uses the SwiGLU activation function over the ReLU, GeLU or even Swish activation functions (these latter activations are commonly used in Transformer architectures).</li></ol><p>2. Parallelised Transformer layer that differs from normal parallelisation</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*AHaPDnGi5Y7TGgw0IKNbJw.jpeg" /></figure><p>The transformer model has two main modules: Attention module and MLP module. In general, the attention layer usually precedes the MLP module. However, in PaLM, the attention and MLP layers are combined together for an overall more compute-efficient architecture.</p><p>As shown in Fig. 1, the normal Transformer layer can be represented as follows</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/710/0*rI6NuJ9ow6Ek6EhI" /></figure><p>While detailed as Fig. 2, the Transformer layer of PaLM can then represented as</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/729/0*y9wSuQFaOIdtjyzS" /></figure><p>According to the paper, fusing the first linear layer of the MLP and the first linear Attention layer can bring an improvement of about 15%! However, we found that even the second layers can be fused, granting an even greater improvement. The architecture of the model after fusion can thus be represented in the following diagram.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/797/0*17_hlZMDXYhOZP9p" /><figcaption>Figure 2: Structure of PaLM Transformer layer</figcaption></figure><p>3. Multi-Query attention mechanism</p><p>Unlike conventional multi-head attention, PaLM has only one head for both the keys and values, while the queries maintain multiple heads as normal multi-head attention. This reduces the number of parameters and thus increases both the speed of training and inference, whilst retaining great model performance on accuracy. We provide both multi-query and multi-head mechanisms in our implementation.</p><p>4. Both linear and layernorm layers do not use bias weights, which researchers from Google claim is beneficial for training stability.</p><p>We firstly reproduced the PaLM model architecture on one GPU according to the PaLM paper’s description. Here we have referred following repo for the reproduction:<a href="https://github.com/lucidrains/PaLM-pytorch"> https://github.com/lucidrains/PaLM-pytorch</a></p><p><strong>ColossalAI’s Enhanced Parallelisation</strong></p><p>Colossal-AI comes to help users easily implement the PaLM model in a parallel manner. Let us describe some of the interesting features that Colossal-AI permits us to write a parallelised, distributed PaLM.</p><p><strong>Tensor parallel</strong></p><p>One of the core tenets of Colossal-AI is ease of use. That’s why we built a system that is consistent with PyTorch’s interface and does not require deep learning practitioners to learn a completely new framework (yet again). For example, simply replacing toch.nn.Linear with colossalai.nn.Linear will enable users to enjoy all of Colossal-AI’s sota HPC techniques (such as tensor parallelism).</p><p>More information can be found here: <a href="https://www.colossalai.org/docs/features/1D_tensor_parallel/">https://www.colossalai.org/docs/features/1D_tensor_parallel/</a></p><p>Tensor parallelism afforded in an intuitive easy to use manner by Colossal-AI is important, particularly in the quest to implement an efficient Pathways model.</p><p>Due to PaLM’s special setting in its attention structure, Colossal-AI needs to deal with a problem involving query, key, and value(Q, K, V). Existing Tensor parallelism in Colossal-AI cuts the last dimension of query, key and value (the first dimension may be cut depending on the parallel mode, but it does not affect the computation, so it will be ignored below), and since key and value are single-head, we need to perform additional communication to ensure their correctness. Here we use B to denote batch size, S to denote sequence length, H to denote hidden size, N to denote the number of attention heads, A to denote the size of a single attention head, and P to denote the number cuts of tensor parallel, where H = NA.</p><p>In the non-parallel case, the size of our multi-head Q is (B, S, H) and the size of single-head K and V is (B, S, A). By converting Q into (B, S, N, A), we can directly compute attention together with K and V. But in the parallel situation, Q is (B, S, H/P) and K and V are (B, S, A/P). We can change Q into (B, S, N/P, A) so that we can cut the head dimension of the Q on different GPUs. But this is still not computable because the values on K and V are not sufficient to form a complete attention head, so we need to introduce an additional all-gather operation to form a completed head, i.e., (B, S, A/P) -&gt; (B, S, A). In this way, normal attention computation can be performed.</p><p><strong>ZeRO Parallelism</strong></p><p>Colossal-AI can additionally provide a memory optimisation by removing redundant operations in the data-parallel approach, dubbed ZeRO parallelism (initially proposed by Microsoft). We can combine the ZeRO parallel approach with different forms of tensor parallelisms as mentioned above to enjoy further gains in efficiency.</p><p><strong>Heterogeneous Training</strong></p><p>To support large-scale AI model training on a single node, we implement a dynamic heterogeneous memory management mechanism. We optimize the placement of tensors on CPUs and GPUs so as to minimize movement of large tensors. In this manner, we thus leverage heterogeneous memory efficiently.</p><p><strong>Training process</strong></p><p>The training process is as simple as it gets with Colossal-AI. One simple configuration file can specify what parallel approaches we require. In the PaLM training process, we specify tensor parallelism as well as the configuration of ZeRO.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*pB1B3KJAb03BgS3b" /></figure><p>With this configuration file setup, we are good to go! We can use a simple API: colossal.initialise, to build our training engine and provide us with APIs that resemble PyTorch style code. We can use Colossal-AI very easily in this manner for large-scale training without sacrificing performance.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*7F6bFUmY9E4aL-p6" /></figure><p><strong>Performance Testing</strong></p><p>We tested all our results on a single server with 8 A100 40GB GPUs. Our network consists of NVLink to interconnect adjacent pairs of GPUs at high speed, and PCI-E between four pairs of GPUs.</p><p>We constructed a PaLM structured network with 8 billion parameters and trained it using a hybrid parallel strategy (a mix of 1D, 2D, 2.5D Tensor Parallel approaches as well as the ZeRO optimization). Colossal-AI can additionally switch between training strategies inexpensively by simply changing the comprehensible configuration file.</p><p>In the following figure, b denotes the batch size of each data parallel process group, XXtpY denotes the tensor-parallel strategy, XX denotes the 1D, 2D, 2.5D parallel scheme, and Y denotes the parallel degree of tensor parallelism. zeroX denotes in ZeRO parallelism setting with a ZeRO parallelism of degree X. Here we have data parallel degree times model parallel degree quals to the total number of GPUs.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/750/0*Y6zaBckyy4ht_Lfd" /></figure><p>Through numerous experiments, we found that heterogeneous training is a necessity. All the aforementioned schemes enjoyed greater compute efficiencies when implemented with heterogeneous training. Without it, running a model with 8 billion parameters would not be possible.</p><p>Specifically, in the cases of 2,4 and 8 GPUs, we find that a 1D tensor parallel degree of 2 works best. This manifests in our setup because the communication bandwidth between adjacent GPUs is high. With the aforementioned optimal configuration, most of the communication is placed between adjacent GPUs. With different network configurations, 2D or even 2.5D tensor parallelism could be more apt. However, with colossal-AI, all that needs to be changed is a simple configuration file, and training can be adapted to a new parallel configuration.</p><p><strong>Summary</strong></p><p>To conclude, we would like to repeat that we have reproduced the architecture of PaLM. Unfortunately, due to the scarcity of computational resources, we could not fully reproduce the full 100 billion parameter model. Moreover, due to PaLM not being open-sourced, our implementation could deviate slightly from the original implementation of Google’s.</p><p>If you have any questions, please feel free to raise an issue or a PR on Github and we will try our best to answer them as soon as possible!</p><p><strong>Funding</strong></p><p>HPC-AI Tech raised 4.7 million USD from top VC firms in just 3 months after the company was founded. For more information, please email <a href="mailto:contact@hpcaitech.com">contact@hpcaitech.com</a>.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=6362c1144e53" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Train 18-billion-parameter GPT models with a single GPU on your personal computer!]]></title>
            <link>https://medium.com/@hpcaitech/train-18-billion-parameter-gpt-models-with-a-single-gpu-on-your-personal-computer-8793d08332dc?source=rss-ec86827abf4------2</link>
            <guid isPermaLink="false">https://medium.com/p/8793d08332dc</guid>
            <category><![CDATA[parallel-computing]]></category>
            <category><![CDATA[distributed-systems]]></category>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[ai]]></category>
            <dc:creator><![CDATA[HPC-AI Tech]]></dc:creator>
            <pubDate>Mon, 16 May 2022 23:51:50 GMT</pubDate>
            <atom:updated>2023-03-07T09:12:05.929Z</atom:updated>
            <content:encoded><![CDATA[<h3>Train 18-billion-parameter GPT models with a single GPU on your personal computer! Open source project Colossal-AI has added new features！</h3><p>When it comes to training large AI models, people will think about using thousands of GPUs, expensive training costs, and only a few tech giants can afford them. While AI users, like researchers from startups or universities, could do nothing but get overwhelmed by news about large models~</p><p><strong>Now, a PC with only one GPU can train GPT with up to 18 billion parameters, and a laptop can also train a model with more than one billion parameters. Compared with the existing mainstream solutions, the parameter capacity can be increased by more than ten times!</strong></p><p>Such a significant improvement comes from <a href="https://github.com/hpcaitech/ColossalAI"><strong>Colossal-AI</strong></a>, which is an efficient training system for general large AI models. Best of all, it’s completely open-sourced and requires only minimal modifications to allow existing deep learning projects to be trained with much larger models on a single consumer-grade graphics card, allowing everyone to train large AI models at home! In particular,<strong> it makes downstream tasks and application deployments such as large AI model fine-tuning and inference much easier!</strong></p><p>By providing various popular efficient parallelisms, <a href="https://github.com/hpcaitech/ColossalAI">Colossal-AI</a> could also help its users easily deploy existing projects to large-scale computing clusters.</p><p>Check out the project over here: <a href="https://github.com/hpcaitech/ColossalAI">https://github.com/hpcaitech/ColossalAI</a></p><h3><strong>Tech giants strive for large AI models</strong></h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Y6N5OU3OZ3Gp5YlaIJ8C7A.png" /></figure><p>After Google proposed the BERT model with 300 million parameters in 2018, the large models’ parameter records have been updated many times in just a few years, such as GPT-3 with 175 billion parameters proposed by OpenAI, MT-NLG with 530 billion parameters introduced by Microsoft and NVIDIA jointly…</p><p>The dense model has achieved the scale with more than hundreds of billions of parameters, while the sparse Mixture of Experts (MoE) model, such as the Switch Transformer released by Google in 2021, has brought the number of parameters to the level of trillion.</p><p>However, training such large models from scratch might be extremely expensive. It usually requires hundreds or even thousands of professional high-performance GPUs such as NVIDIA A100 at the same time. If we use a dedicated InfiniBand high-speed network to build a supercomputer cluster, the cost for training could even reach ten million dollars.</p><h3><strong>Use a single consumer-level GPU to train large AI models</strong></h3><p>It’s obvious that AI users like university students and individual developers could not afford such high costs to train large models, and the most popular computing resources are the NVIDIA RTX GPUs for this kind of people in the AI community.</p><p>In order to enhance AI productivity, allow large models to benefit more developers, and truly realize our vision to make the use of large AI models “fast and cheap”, <strong>Colossal-AI requires only a few lines of codes to achieve a ten-fold increase in the capacity of model training</strong>.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/942/1*EEyMJG-U9tEHvrSR1XuTRA.png" /></figure><p>On all types of hardware, Colossal-AI performs better than vanilla PyTorch and mainstream distributed solutions such as Microsoft’s DeepSpeed.</p><p>For the representative of large models — GPT, Colossal-AI is capable of training it with up to 1.5 billion parameters on a gaming laptop with RTX 2060 6GB. For a PC with RTX3090 24GB, Colossal-AI could help its users to train GPT with 18 billion parameters. For high performance graphics cards such as Tesla V100, Colossal-AI could bring significant improvements as well.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*XiWXXfU_wcTvVemUvmRSig.png" /></figure><p>Colossal-AI has also successfully implemented Google’s PaLM (Pathways Language Model), which was published recently. It has also shown excellent performance improvements on various hardware, while Microsoft DeepSpeed has not published its official PaLM implementation.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*ygsxhvL_JagpdmUvxnmtcA.png" /></figure><h3><strong>Key technology: Enhanced heterogeneous training</strong></h3><p>The biggest problem with using a single consumer-grade GPU to train a large AI model is that the GPU memory capacity is extremely limited, which severely restricts the model parameters that can be accommodated. The ZeRO-offload method proposed by Microsoft DeepSpeed tries to split the model and utilize the CPU memory with larger capacity and lower cost. At present, there have been several modified versions based on DeepSpeed for heterogeneous training. But as shown on the left part of the figure below, when the GPU memory is insufficient for its corresponding model requirements, the system will crash even if there is still memory available on the CPU.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Pcaic_oHvRfp0_ObFHYevA.png" /></figure><p>Different from the derivatives based on DeepSpeed, Colossal-AI team built its core technologies such as ZeRO from scratch, solving current problems like DeepSpeed only statically divides model data between CPU and GPU memory, and only uses fixed memory layout for different training configurations. A lot of improvements have been made by Colossal-AI to improve the efficiency of the usage of GPU and CPU memory. After all, CPU memory is much cheaper than high performance graphics cards with large memory.</p><p>The Gemini mechanism designed by Colossal-AI, <strong>efficiently manages and utilizes the heterogeneous memory of GPU and CPU</strong>, so that tensors are dynamically distributed in the storage space of CPU-GPU during the training process, therefore model training can break the GPU’s memory barrier.</p><p>We take advantage of the iterative nature of the deep learning network training process, and divide the training into two stages: warmup stage and non-warmup stage according to the number of iterations. In the initial warmup phase, memory information is monitored; in the non-warmup phase, the collected information is used to efficiently move tensors to minimize CPU-GPU data movement.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*uA5UCmg6h5YiXlUXCo9anA.png" /></figure><p>It sounds easy, but its implementation is troublesome. It is actually hard to obtain the memory usage of non-model data, because the life cycle of non-model data is not managed by users, and the existing deep learning frameworks do not expose the tracking interface of non-model data to its users. Secondly, non-framework overhead such as CUDA context also needs to be considered.</p><p>Colossal-AI obtains the usage of CPU and GPU memory by sampling in the warmup stage. While the usage of non-model data can be obtained by comparing the maximum system memory usage and model memory usage between two moments. The memory usage of the model could be known by querying the memory manager, as shown by the black solid line in the figure below.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*JZT1eWQUK4JB8NBF0g8fOw.png" /></figure><p>All tensors from models are managed by the memory manager, and each tensor is marked with information of states, including HOLD, COMPUTE, FREE, etc. According to the dynamically queried memory usage, Colossal-AI changes states of tensors, and adjusts positions of tensors continuously. Finally, the efficient usage of GPU and CPU memory is realized, maximizing model capacity and balancing training speed in the case of extremely limited hardware, which is of <strong>great significance for AI democratization and low-cost fine-tuning of downstream tasks for large models</strong>.</p><h3><strong>Furthermore: convenient and efficient parallelizations</strong></h3><p>Parallel and distributed technologies are important methods to further accelerate model training. To train the world’s largest and most advanced AI models within the shortest time, efficient distributed parallelization is still a necessity. Aiming at the pain points of the existing solutions such as limited parallel dimension, low efficiency, poor versatility, difficult deployment, and lack of maintenance, Colossal-AI uses technologies such as efficient multi-dimensional parallelism and heterogeneous parallelism to <strong>allow users to deploy large AI models efficiently and quickly with only a few modifications of their codes</strong>.</p><p>For example, for a super-large AI model such as GPT-3, compared to the NVIDIA solution, Colossal-AI <strong>only needs half the computing resources </strong>to start training; if the same computing resources are used, the speed could be further increased by 11%, which could <strong>reduce the training cost of GPT-3 over a million dollars</strong>.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/977/1*MhslE3-o2uS6GRy-EZgVKw.png" /></figure><p>For AlphaFold, which is used for protein structure prediction, our team has introduced FastFold based on the Colossal-AI acceleration scheme. FastFold successfully surpassed other schemes proposed by Google and Columbia University, reducing the training time of AlphaFold from 11 days to 67 hours, and the total cost is lowered as well. Also, in long sequence inference, we achieved a speed improvement of 9.3 ~ 11.6 times.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/977/1*A9wLN0RYtiMnUHTzcFdefQ.png" /></figure><p>Besides, Colossal-AI values open source community construction, providing English and Chinese tutorials, and supporting the latest cutting-edge applications such as PaLM and AlphaFold. Colossal-AI will roll out new and innovative features regularly as well. We always welcome suggestions and discussions from the community, and we would be more than willing to help you if you encounter any issues. You can raise an<a href="https://github.com/hpcaitech/ColossalAI/issues"> issue</a> here or create a discussion topic in our <a href="https://github.com/hpcaitech/ColossalAI/discussions">forum</a>. Your suggestions are highly appreciated here. Recently, <a href="https://github.com/hpcaitech/ColossalAI">Colossal-AI</a> reached <strong>No. 1 on the top trending projects on Github</strong>, against a backdrop of many projects that have as many as 10K stars.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/977/1*Epbz0Y-nXNRxr9OrdjkyOw.png" /></figure><h3><strong>About us</strong></h3><p>The core members of the <a href="https://hpcaitech.com">HPC-AI Tech</a> team are from the UC Berkeley, Stanford University, Tsinghua University, Peking University, National University of Singapore, Nanyang Technological University and other well-known universities. Also, they have work experience from tech giants like Google Brain, IBM, Intel, Microsoft, NVIDIA, etc. The company has received seed round funds from many top VC institutions such as Sinovation Ventures and ZhenFund as well.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/772/1*UIx3nqth_D9mvrWpXGNLmQ.png" /></figure><p><em>Prof. Yang You, Founder of HPC-AI Tech</em></p><p><em>Ph.D., University of California, Berkeley</em></p><p><em>IPDPS/ICPP Best Paper Author</em></p><p><em>ACM/IEEE CS George Michael Memorial HPC Fellowship</em></p><p><em>Forbes 30 Under 30 (Asia 2021)</em></p><p><em>IEEE-CS Outstanding Newcomer Award in Supercomputing</em></p><p><em>UC Berkeley EECS Lotfi A. Zadeh Prize</em></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/977/1*xvgCkbeT5gz42gm7siof7Q.png" /></figure><p><em>Prof. James Demmel, CSO of HPC-AI Tech</em></p><p><em>Distinguished Professor, University of California, Berkeley</em></p><p><em>ACM/IEEE Fellow</em></p><p><em>Member of the American Academy of Sciences, the Academy of Engineering, and the Academy of Arts and Sciences</em></p><h3><strong>Portal</strong></h3><p>Code:</p><p><a href="https://github.com/hpcaitech/ColossalAI">GitHub - hpcaitech/ColossalAI: Making large AI models cheaper, faster and more accessible</a></p><p>Reference:</p><p><a href="https://www.deepspeed.ai">Latest News</a></p><p>Fang, Jiarui, et al. “Parallel Training of Pre-Trained Models via Chunk-Based Dynamic Memory Management.” <em>IEEE Transactions on Parallel and Distributed Systems</em> 34.1 (2022): 304–315.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=8793d08332dc" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[5 Must-Follow Features That Are Seeing Colossal-AI’s Success]]></title>
            <link>https://medium.com/@hpcaitech/5-must-follow-features-that-are-seeing-colossal-ais-success-2d5361e27e4b?source=rss-ec86827abf4------2</link>
            <guid isPermaLink="false">https://medium.com/p/2d5361e27e4b</guid>
            <dc:creator><![CDATA[HPC-AI Tech]]></dc:creator>
            <pubDate>Tue, 05 Apr 2022 12:59:36 GMT</pubDate>
            <atom:updated>2022-04-07T02:16:21.168Z</atom:updated>
            <content:encoded><![CDATA[<p>Colossal-AI, HPC-AI Tech’s flagship large-scale parallel AI training system, has become an overnight success. Recently, it was touted as one of the top trending projects on Github, against a backdrop of many projects that have as many as 10K stars.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*BheRKCvQSon1oB5z" /></figure><p>Colossal-AI is an easy-to-use deep learning system that enables users to maximize the efficiency of AI deployments whilst drastically reducing costs. If you would like to learn more about the project, do check out our GitHub repository: <a href="https://github.com/hpcaitech/ColossalAI">https://github.com/hpcaitech/ColossalAI</a></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/989/0*uI-01zh9dvJSG593" /></figure><p>In celebration of the first official release of Colossal-AI, we describe some of its awesome new features that we believe are responsible for its success:</p><ul><li>A brand new ZeRO module;</li><li>Profiler TensorBoard plugin of Beta version (Finer grained monitoring than PyTorch for communication, memory, etc);</li><li>MOE feature and example;</li><li>Fix bugs and complete tutorials;</li><li>Making Colossal-AI work with HuggingFace;</li></ul><p>As well as some of its core features that has made Colossal-AI a popular choice in training and deploying large AI models.</p><h3><strong>Professional Help with Large Model Training</strong></h3><p>Firstly, it’s important to note that Colossal-AI solves an <em>important</em> problem. In recent years, deep learning has seen an insurgence of large models dominating performance charts. AI models that operate at the frontier have increased ten thousand-fold in just a few years, far outpacing the growth of hardware technologies. Not only are these large AI models far beyond the capacity of a single GPU, but they also often demand hundreds of years of a single GPU.</p><p>It thus becomes imperative to improve the capacity of a single GPU via its efficient use to achieve high-performance training of large AI models. This is precisely what Colossal-AI does <em>without burdening the programmer</em>.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*0TJwgyIMR8BFELenLabRzA.png" /></figure><p>Colossal-AI empowers the developer to write performant code for AI deployments through a variety of techniques such as multi-dimensional parallelism and better tools for the maintenance and deployment of models. It allows the programmer to both quickly and efficiently deploy large AI models as well as train them in a distributed manner with only minimal code level modifications. Colossal-AI’s claim to fame are its efficient multi-dimensional parallelism, GPU memory optimization, large-scale optimizer library and fine-grained monitoring.</p><p>Let’s dive into some of these key features to see what makes Colossal-AI really great!</p><h4><strong>[Key Features] Multi-dimensional Parallelism</strong></h4><p>Existing solutions apply limited modes of parallelism in the process of training large scale AI-models, primarily: data parallelism, one-dimensional tensor parallelism and pipeline parallelism. Colossal-AI, however, provides further modes of parallelism, including 2, 2.5, 3-dimensional tensor parallelism as well as sequence parallelism. It offers additional multidimensional hybrid parallel solutions on top of these as well.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*Yawca62I3W3J22fG" /><figcaption>14x batch size and 5x training speed when tensor parallelism of ViT is 64</figcaption></figure><p>Amongst these new parallel techniques, high-dimensional tensor parallelism can greatly reduce memory consumption, improve communication efficiency, and make more efficient use of computing resources.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*HGPsY1UFMg-r2abQ" /><figcaption>Sequence parallelism helps BERT improve training speed by 2X, or 1.5X sequence length</figcaption></figure><p>Colossal-AI even engineers a new form of parallelism: sequence parallelism, which can also be applied to large images, video, long text, long periods of medical surveillance and other data. It enables machines to process long pieces of sequential data that were previously not possible.</p><h4><strong>[New Features] GPU Memory Optimization</strong></h4><p>Next up is Colossal-AI’s GPU memory optimization innovations. It combines multiple GPU memory optimization technologies, including multi-dimensional parallelism, ZeRO redundancy memory elimination, CPU offloading, Gradient Checkpointing, Automatic Mixing Accuracy (AMP), etc., to help users avoid memory bottlenecks and to reduce training resource requirements.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*bB8TF7VYRXWiP0kO" /><figcaption>GPT-2 uses Colossal-AI to train a model 24 times the size, or 3 times acceleration with the same hardware</figcaption></figure><h4><strong>[Product Improvement &amp; New Features] Ease of Use</strong></h4><p>Finally, Colossal-AI prizes simplicity and ease of use. It is designed to be compatible with PyTorch, allowing existing projects to work with minimal modifications. In addition, the system is easily extensible, making it easy to add new features as needed whilst maintaining its performance.</p><p>It adds a new<strong> fine-grained monitoring</strong> tool that allows developers to monitor the network, communication, and memory states within an iteration. Compared to existing frameworks, which can only record the training process with iterations, Colossal-AI makes it easy for developers to accurately analyze and debug deep learning code.</p><p>Lastly, Colossal-AI provides a <strong>large-scale optimizer</strong> <strong>library </strong>including efficient optimizers like LAMB and LARS, which extend the training batch size to 65536 for the first time. It is also compatible with all of PyTorch’s own optimizers. It is now easier than ever to use large batch optimization methods to train large AI models with Colossal-AI.</p><h4><strong>[Benefits] Solutions for Industry</strong></h4><p>This is all great in theory, but what about in practice? Luckily, Colossal-AI has proven its capabilities in its application to hard problems across a variety of industries such as autonomous driving, cloud computing, retail, medicine and chip production. It has also established cooperation with top open source AI organizations such as Hugging Face.</p><ul><li><strong>Helping Drug Research and Development: FastFold</strong></li></ul><p>One such monumental application is to the domain of protein folding. Recently, AlphaFold was selected by Science and Nature as one of the top 10 scientific breakthroughs in 2021 for its powerful AI ability to predict protein structure. Nonetheless, it suffers from lengthy training times and high costs.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*ru7gcu_JJ7bo64_X" /></figure><p>We applied Colossal-AI to develop an accelerated AI model to predict protein structures: FastFold. It introduces Colossal-AI’s novel GPU optimization and large-model training techniques to AlphaFold training and inference. Fastfold, successfully outperforms the solutions from Google and Columbia University, reducing AlphaFold training time <strong>from 11 days to 67 hours at a lower total cost whilst achieving a 9.3~11.6x speedup</strong> in long-sequence inference.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*W97sNt93r-s66_r0" /><figcaption>Comparison of inference performance for long sequences</figcaption></figure><ul><li><strong>GPT-3 Training with Half of the Machines</strong></li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*CK62i-FlDAky47O3" /></figure><p>For veteran state-of-the art models, Colossal-AI yet again makes headways. For the notorious GPT-3 model, Colossal-AI requires only half the computational resources to start training compared with NVIDIA’s Megatron-LM. This means that if we use the same computational resources, the speedup incurred via Colossal-AI amounts to ~11%, which can reduce the cost to train GPT-3 by over a million dollars!</p><h3>Your Thoughts?</h3><p>Colossal-AI will roll out new and innovative features regularly. We always welcome suggestions and discussions from the community, and we would be more than willing to help you if you encounter any issue. You can raise an<a href="https://github.com/hpcaitech/ColossalAI/issues"> issue</a> here or create a discussion topic in the<a href="https://github.com/hpcaitech/ColossalAI/discussions"> forum</a>. Your voice is a perfect tutor to Colossal-AI.</p><h3>Portal</h3><p>Code: <a href="https://github.com/hpcaitech/ColossalAI">https://github.com/hpcaitech/ColossalAI</a></p><p>Paper: <a href="https://arxiv.org/abs/2110.14883">https://arxiv.org/abs/2110.14883</a></p><p>Tutorial: <a href="https://www.colossalai.org/">https://www.colossalai.org/</a></p><h3><strong>Join Us</strong></h3><p><a href="https://www.hpcaitech.com/">HPC-AI Tech</a> is a global team and the core members are from the University of California, Berkeley, Stanford University, Tsinghua University, Peking University, National University of Singapore, Singapore Nanyang Technological University, and other top universities in the world. Currently, HPC-AI Tech is recruiting full-time/intern AI system/architecture/compiler/network/CUDA/SaaS/k8s core system developers, open source program operators, and sales personnel.</p><p>HPC-AI Tech provides highly competitive compensation packages. Our staff can also work remotely. You are also welcome to recommend outstanding talents to HPC-AI Tech. If they successfully join HPC-AI Tech, we will provide you with a recommendation fee of thousands of US dollars.</p><p>Resume delivery mailbox: <a href="mailto:hr@hpcaitech.com">hr@hpcaitech.com</a></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/494/0*tWybKqr1l4NV3K7N" /></figure><blockquote>Prof. Yang You, Founder of HPC-AI Tech</blockquote><blockquote>Ph.D., University of California, Berkeley</blockquote><blockquote>IPDPS/ICPP Best Paper</blockquote><blockquote>ACM/IEEE George Michael HPC Fellowship</blockquote><blockquote>Forbes Elite Under 30 (Asia 2021)</blockquote><blockquote>IEEE-CS Outstanding Newcomer Award in Supercomputing</blockquote><blockquote>UC Berkeley EECS Lotfi A. Zadeh Outstanding Graduate Award</blockquote><figure><img alt="" src="https://cdn-images-1.medium.com/max/750/0*MMwWC-VHqpAZjuRb" /></figure><blockquote>Prof. James Demmel, CSO of HPC-AI Tech</blockquote><blockquote>Distinguished Professor, University of California, Berkeley</blockquote><blockquote>ACM/IEEE Fellow</blockquote><blockquote>Member of the American Academy of Sciences, the Academy of Engineering, and the Academy of Arts and Sciences</blockquote><h3><strong>Funding</strong></h3><p><strong>HPC-AI Tech raised 4.7 million USD from top VC firms in just 3 months after the company was founded. For more information, please email contact@hpcaitech.com</strong></p><h3>Reference</h3><p><a href="https://unsplash.com/photos/4YoINz4XvnQ?utm_source=unsplash&amp;utm_medium=referral&amp;utm_content=creditShareLink">https://unsplash.com/photos/4YoINz4XvnQ?utm_source=unsplash&amp;utm_medium=referral&amp;utm_content=creditShareLink</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=2d5361e27e4b" width="1" height="1" alt="">]]></content:encoded>
        </item>
    </channel>
</rss>