AI Inference | 人工智慧推論

  • What is it?
    AI inference is the second step in the two-part process that makes up machine learning and deep learning; the first step is AI training. The two steps are an important reason why modern artificial intelligence is suitable for such a diverse range of tasks, from generating content to driving autonomous vehicles.

    During the inference phase, a pre-trained AI model is exposed to fresh unlabeled data. It relies on the database that it "studied" during its training to analyze the new input and respond with the correct output. To use generative AI as an example, every time you ask ChatGPT a question, or ask Stable Diffusion to draw you something, the AI model is inferencing. The reason it can come up with such human-like responses is because of all the training that it went through before.

    Even as it engages in inferencing, the AI is also recording the responses from human users for its next training session. It takes note when its output is praised or criticized. In this way, the continuous loop of training and inference makes AI more and more lifelike.

  • Why do you need it?
    The whole reason we train AI models is so that they can inference—interact with new data in the real world and help humans lead more productive and comfortable lives. A lot of what advanced AI products can do for us, from reading human handwriting to recognizing human faces, from piloting driverless vehicles to generating content, is AI inference at work. When you hear terms like computer visionnatural language processing (NLP), or recommendation systems—these are all instances of AI inference.

  • How is GIGABYTE helpful?
    To conduct AI inference efficiently, you need a computing platform with good processing speeds and the low latency to match. The reason is simple: the AI model will likely be servicing a lot of users at the same time. Especially in scenarios where a speedy response may affect productivity (such as sorting mail in a distribution center) or even safety (such as controlling a self-driving car), attributes like high performance and low latency become even more pertinent.

    On the server side, one of the best solutions for AI inference is GIGABYTE Technology's G293-Z43, which boasts an industry-leading configuration of 16 AMD Alveo™ V70 cards in a 2U chassis. The Alveo™ V70 accelerator is based on AMD’s XDNA™ architecture, which is optimized for AI inference. The Qualcomm® Cloud AI 100 solution is another product that can help data centers engage in inferencing on the cloud and the edge more effectively, due to its advancements in signal processing, power efficiency, node advancement, and scalability.

    Within individual vertical markets, GIGABYTE also offers bespoke hardware for different applications. For example, in the automotive and transportation industry, GIGABYTE's Automated Driving Control Unit (ADCU) is an embedded in-vehicle computing platform with AI acceleration; it's been deployed in Taiwan's self-driving buses. For AI-based facial recognition, which has seen broad adoption in the retail and education sectors, GIGABYTE's AI Facial Recognition Solution is an all-in-one solution that can achieve an accuracy level of 99.9% in the 1vN model.

    Learn more : 《Advance AI with GIGABYTE’s supercharged AI server solutions

  • WE RECOMMEND
    RELATED ARTICLES
    如何將人工智慧導入醫療保健業

    AI & AIoT

    如何將人工智慧導入醫療保健業

    從事醫療與健康產業的讀者,請花幾分鐘閱讀本篇文章,了解人工智慧(AI)為這個領域所帶來的嶄新商機,並認識能助您從中受益的AI工具。本篇是技嘉科技「Power of AI」系列文章,目的是介紹不同產業的最新AI趨勢,鼓勵前瞻者把握AI浪潮,找尋自己的機會。
    如何挑選您的AI伺服器?(下)記憶體、儲存裝置和其他元件

    Tech Guide

    如何挑選您的AI伺服器?(下)記憶體、儲存裝置和其他元件

    人工智慧盛行的當下,各種組織積極導入「AI伺服器」。技嘉科技最新發表的《科技指南》:「如何挑選您的AI伺服器?」,文章下半篇將介紹CPU和GPU以外的六個關鍵零組件。挑選合適元件可讓AI伺服器的性能達到顛峰,勝任人工智慧的相關工作。
    如何挑選您的AI伺服器?(上)CPU和GPU

    Tech Guide

    如何挑選您的AI伺服器?(上)CPU和GPU

    生成式人工智慧和其他AI工具盛行的當下,挑選合適的AI伺服器成為各產業的首要任務。技嘉科技最新發表《科技指南》,帶領讀者認識AI伺服器的八個關鍵零組件,本篇從最重要的元件開始,即中央處理器(CPU)和圖形處理器(GPU)。挑選適當的運算晶片,打造量身訂做的人工智慧超算平台,可以讓工作事半功倍,為使用者開創全新的巔峰。
    帶您快速跟上人工智慧AI趨勢的十大問答

    AI & AIoT

    帶您快速跟上人工智慧AI趨勢的十大問答

    大家都在談人工智慧(AI),您是否也希望擁有基本的知識,參與這個話題的討論?別擔心,技嘉科技為您準備了介紹AI趨勢的十大問答,讓您能快速理解人工智慧的概念!
    如何將人工智慧導入汽車和運輸產業

    AI & AIoT

    如何將人工智慧導入汽車和運輸產業

    若您從事汽車與運輸業,請花幾分鐘閱讀本篇文章,了解人工智慧(AI)所開拓的嶄新機會,認識能助您拓展更多可能性與商機的科技工具。本篇是技嘉科技「Power of AI」系列文章,目的是介紹不同產業的AI趨勢,協助具有先見之明的前瞻者利用AI創造自己的機會。
    如何挑選適當的伺服器冷卻方案? 技嘉科技《科技指南》系列文章

    Tech Guide

    如何挑選適當的伺服器冷卻方案? 技嘉科技《科技指南》系列文章

    隨著科技進步,新一代的處理器使用更多電力、產出更多熱能。選購伺服器時,應當留意溫控問題,好的冷卻方案可確保伺服器正常運作,且不至於太耗電、或是需要頻繁的維修。 技嘉科技是高性能伺服器的領導品牌,本篇《科技指南》針對市面上廣泛使用的三種散熱方法(氣冷式、液冷式和浸沒式)逐一說明,並介紹技嘉的相關產品,協助您挑選最適合的解決方案。
    揭秘生成式AI:從「訓練」到「推論」,一窺其神奇運作原理

    Tech Guide

    揭秘生成式AI:從「訓練」到「推論」,一窺其神奇運作原理

    生成式人工智慧(generative AI)來勢洶洶,實用案例包括聊天機器人ChatGPT能撰寫文案與企劃案、Stable Diffusion等圖像生成工具能依指令產出栩栩如生的影像。事實上,想搞懂生成式AI不難,使用它比想像中還容易。技嘉科技在本篇《科技指南》中,為您解剖生成式AI背後「訓練」與「推論」兩個重要步驟,並為您推薦專為人工智慧開發所設計的技嘉解決方案,協助您成為AI世代的佼佼者。
    技嘉伺服器加持 國研院國網中心與冉色斯引導台灣動畫走向國際

    Success Case

    技嘉伺服器加持 國研院國網中心與冉色斯引導台灣動畫走向國際

    瀚草影視製作影集《2049》,2021年在網飛、台視、東森等平台播出;冉色斯動畫新作《2049+絕處逢聲》,是影集《2049》衍生系列,同步在myVideo首播。精妙絕倫的電腦動畫,使用國研院國網中心算圖農場的技嘉科技高效能運算伺服器渲染製作,透過伺服器搭載的NVIDIA®加速卡,動畫師能駕馭頂尖渲染技術,平行運算功能讓多名用戶能同時使用算圖服務,智慧管理功能則確保服務穩定性。千呼萬喚的台灣卡通巨作問世,《2049+絕處逢聲》受到國內外串流平台高度關注,台灣科技文創業的創意與巧思,期望能被全世界看見。