{"id":144637,"date":"2023-08-09T15:30:06","date_gmt":"2023-08-09T06:30:06","guid":{"rendered":"http:\/\/ee.presscat.kr\/?post_type=ai-in-circuit&#038;p=144637"},"modified":"2026-04-05T10:46:44","modified_gmt":"2026-04-05T01:46:44","slug":"ee-professor-kim-joo-young-developed-a-chatgpt-core-ai-semiconductor-with-a-2-4-fold-improvement-in-price-efficiency","status":"publish","type":"ai-in-circuit","link":"http:\/\/ee.presscat.kr\/en\/ai-in-circuit\/ee-professor-kim-joo-young-developed-a-chatgpt-core-ai-semiconductor-with-a-2-4-fold-improvement-in-price-efficiency\/","title":{"rendered":"EE Professor Kim Joo-Young Developed A ChatGPT Core AI Semiconductor with A 2.4-fold Improvement in Price efficiency"},"content":{"rendered":"<p><span style=\"font-size: 14pt\"><strong><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">EE Professor Kim Joo-Young Developed A ChatGPT Core AI Semiconductor with A 2.4-fold Improvement in Price efficiency<\/span><\/strong><\/span><\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<p><img fetchpriority=\"high\" decoding=\"async\" class=\"alignnone size-full wp-image-144604\" src=\"http:\/\/ee.presscat.kr\/wp-content\/uploads\/2023\/08\/\ud558\uc774\ud37c\uc5d1\uc140.jpg\" alt=\"\" width=\"1542\" height=\"302\" title=\"\"><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">The ChatGPT released by OpenAI has captured global attention, and everyone is closely observing the changes this technology will bring out. <\/span><\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">This technology is based on large language models (LLM), which represent an unprecedented scale of artificial intelligence (AI) models compared to conventional AI. <\/span><\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">However, the operation of these models requires a significant number of high-performance GPUs, leading to astronomical computing costs. \u00a0<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">KAIST (President: Lee Kwang-Hyung) announced that research team led by EE Professor Kim Joo-Young Kim has successfully developed an AI semiconductor that efficiently accelerates the inference operations of large language models, which play a crucial role in ChatGPT.\u00a0<\/span><\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">The developed AI semiconductor, named the &#8216;Latency Processing Unit (LPU),&#8217; efficiently accelerates the inference operations of large language models. It incorporates a high-speed computing engine capable of maximizing memory bandwidth utilization and performing all necessary inference computations rapidly.<\/span><\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\"> Additionally, it comes equipped built-in networking capabilities, making it easily expandable with multiple accelerators.\u00a0This LPU-based acceleration appliance server achieved up to a 50% higher performance and approximately 2.4 times better performance-to-price ratio compared to a supercomputer based on the industry-leading high-performance GPU, NVIDIA A100.<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">This advancement holds the potential to replace high-performance GPUs in data centers that are experiencing a rapid surge in demand for generative AI services.\u00a0This research was conducted by Professor HyperExcel Co., founded by Professor Kim Joo-Young and achieved the remarkable accomplishment of receiving the &#8220;Engineering Best Presentation Award&#8221; at the International Design Automation Conference (DAC 2023) held in San Francisco on July 12th (U.S. time).<\/span><\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">DAC is a prestigious international conference in the field of semiconductor design, particularly showcasing global semiconductor design technologies related to Electronic Design Automation (EDA) and Semiconductor Intellectual Property (IP).<\/span><\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\"> DAC attracts participation from renowned semiconductor design companies such as Intel, NVIDIA, AMD, Google, Microsoft, Samsung, TSMC, as well as top universities including Harvard, MIT, and Stanford.<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">Among the world\u2019s notable semiconductor technologies, Professor Kim&#8217;s team stands out as the sole recipient of an award for AI semiconductor technology tailored for large language models.<\/span><\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\"> This award acknowledges their AI semiconductor solution as a groundbreaking means to drastically reduce the substantial costs associated with inference operations for large language models on the global stage.<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\"> Professor Kim stated, &#8220;With the new processor &#8216;LPU&#8217; for future large AI computations, I intend to pioneer the global market and take a lead over big tech companies in terms of technological prowess.&#8221;<\/span><\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">(Note: The provided translation is an elaboration and summary of the original text for clarity and readability.)<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\"><img decoding=\"async\" class=\"alignnone size-full wp-image-144602\" src=\"http:\/\/ee.presscat.kr\/wp-content\/uploads\/2023\/08\/\uc5f0\uad6c\ud300\uc0ac\uc9c4-\uc601\ubb38.jpg\" alt=\"\" width=\"1143\" height=\"291\" title=\"\"><\/span><\/p>\n<div><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">[Related News]<\/span><\/div>\n<div><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">Chosun Ilbo :\u00a0<a style=\"color: #000000\" href=\"https:\/\/biz.chosun.com\/science-chosun\/technology\/2023\/08\/04\/AOMZEDZDJFAKTOTC7YEWDCIJYA\/?utm_source=naver&amp;utm_medium=original&amp;utm_campaign=biz\" target=\"_blank\" rel=\"noopener\">\ucc57GPT \uac00\uc131\ube44 2.4\ubc30 \ub192\uc774\ub294 \ubc18\ub3c4\uccb4 \ub098\uc654\ub2e4 &#8211; \uc870\uc120\ube44\uc988 (chosun.com)<\/a>\u00a0<\/span><\/div>\n<div><span style=\"font-family: verdana, geneva, sans-serif;color: #000000\">DongA Science :\u00a0<a style=\"color: #000000\" href=\"https:\/\/www.dongascience.com\/news.php?idx=61064\" target=\"_blank\" rel=\"noopener\">\ucc57GPT \ud6a8\uc728 \ub192\uc77c &#8216;AI \ubc18\ub3c4\uccb4&#8217; \uac1c\ubc1c\u2026\uad6d\uc81c\ud559\ud68c\uc11c &#8216;\ucd5c\uace0 \ubc1c\ud45c\uc0c1&#8217; : \ub3d9\uc544\uc0ac\uc774\uc5b8\uc2a4 (dongascience.com)<\/a>\u00a0<\/span><\/div>\n<div>\u00a0<\/div>\n","protected":false},"excerpt":{"rendered":"<p>1298<\/p>\n","protected":false},"featured_media":0,"template":"","class_list":["post-144637","ai-in-circuit","type-ai-in-circuit","status-publish","hentry"],"acf":[],"_links":{"self":[{"href":"http:\/\/ee.presscat.kr\/en\/wp-json\/wp\/v2\/ai-in-circuit\/144637","targetHints":{"allow":["GET"]}}],"collection":[{"href":"http:\/\/ee.presscat.kr\/en\/wp-json\/wp\/v2\/ai-in-circuit"}],"about":[{"href":"http:\/\/ee.presscat.kr\/en\/wp-json\/wp\/v2\/types\/ai-in-circuit"}],"wp:attachment":[{"href":"http:\/\/ee.presscat.kr\/en\/wp-json\/wp\/v2\/media?parent=144637"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}