{"id":1000002533,"date":"2025-09-10T19:59:26","date_gmt":"2025-09-10T19:59:26","guid":{"rendered":"https:\/\/alifsemi.com\/?p=1000002533"},"modified":"2025-10-02T19:19:23","modified_gmt":"2025-10-02T19:19:23","slug":"anticipation-mounts-as-developers-dream-of-new-use-cases-for-generative-ai-at-the-edge","status":"publish","type":"post","link":"https:\/\/alifsemi.com\/anticipation-mounts-as-developers-dream-of-new-use-cases-for-generative-ai-at-the-edge\/","title":{"rendered":"Anticipation mounts as developers dream of new use cases for generative AI at the edge"},"content":{"rendered":"\n<p>Everyone knows what generative AI is for, don\u2019t they?<\/p>\n\n\n\n<p>When generative AI is performed in a data center, with access to practically unlimited compute and power resources, it is the modern miracle which provides A+ answers to examination questions, generates \u2018photographs\u2019 which look like images taken by a real camera, and predicts molecular behavior that previously would have needed to be observed in a laboratory experiment.<\/p>\n\n\n\n<p>What about generative AI at the edge and endpoint though? The technology is incredibly exciting.<\/p>\n\n\n\n<p>But what exactly is it for?<\/p>\n\n\n\n<p>This question was once of only hypothetical interest, but now, it\u2019s affecting the decisions of embedded device developers for current design projects. That\u2019s because the <a href=\"https:\/\/alifsemi.com\/press-release\/alif-semiconductor-cements-leadership-position-in-edge-ai-space-with-release-of-benchmark-results-from-its-latest-genai-enabled-mcus\/\" target=\"_blank\" rel=\"noreferrer noopener\"><strong><span style=\"text-decoration:underline;color:#5ACBEE;\"><strong>launch<\/strong><\/span><\/strong><\/a> of the new generation of Ensemble microcontrollers and fusion processors has given developers for the first time a hardware platform which supports transformer operations, the key software function which underpins generative AI models.<\/p>\n\n\n\n<p>The low power consumption and fast inference performance of the new Ensemble E4, E6 and E8 products make the implementation of generative AI a real possibility in endpoint devices, even those powered by very small batteries. It\u2019s a big move for the AI market: the large language models which enable generative AI services such as Gemini and Perplexity run on huge arrays of large, expensive, power-hungry GPUs in data centers.<\/p>\n\n\n\n<p><strong>Bringing generative AI within the scope of battery-powered systems<\/strong><\/p>\n\n\n\n<p>Now, generative AI can run locally in small, battery-powered embedded devices for the home, office, factory or smart city \u2013 and it does not need an array of GPUs. Instead, Alif has assembled in the Ensemble devices the features which enable generative AI at the edge:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>An NPU, the Arm\u00ae Ethos\u2122-U85, which can perform transformer operations and other neural networking operations at high speed and low power<\/li>\n\n\n\n<li>Tight integration of the NPU with the CPUs \u2013 Arm Cortex\u00ae-M55 in the dual-core Ensemble E4 and Cortex-M55 plus Cortex-A32 in the E6 and E8 fusion processors<\/li>\n\n\n\n<li>A wide memory subsystem for extremely fast on- and off-chip transactions<\/li>\n\n\n\n<li>Support for up to two MIPI-CSI image sensors, alongside a fully hardware-accelerated image signal processor pipeline which operates at up to 60fps at 2MP resolution, enabling high-speed vision AI processing<\/li>\n<\/ul>\n\n\n\n<p>This hardware platform makes generative AI possible at amazingly low power: just 36mW when executing a small language model (SLM) to generate text on an Ensemble E4, for instance.<\/p>\n\n\n\n<p><strong>Where natural language capability adds value<\/strong><strong><\/strong><\/p>\n\n\n\n<p>So with generative AI now feasible at the edge and endpoint, developers are starting to figure out what to do with it. Some types of devices obviously lend themselves to the use of generative AI: in smart glasses, for instance, real-time translation of a foreign language is a valuable use case, as well as understanding natural language instructions or queries from the user. And in a smart security camera, generative AI could be used to produce context-aware spoken warnings to a potential intruder. (\u2018If you\u2019re the adult male in the black hoodie who has been hanging around for the past five minutes, you should know that I have taken video footage of you and uploaded it to the local police department.\u2019)<\/p>\n\n\n\n<p><strong>Discovering what the developer\u2019s imagination can invent<\/strong><\/p>\n\n\n\n<p>So there are some natural-language functions which can obviously supplement the operation of existing types of products.<\/p>\n\n\n\n<p>But the capabilities of generative AI extend far beyond language. What might the possibilities be in embedded devices?<\/p>\n\n\n\n<p>All that we know at this stage is that many potential uses of generative AI at the edge are yet to be discovered or imagined. This is completely natural, given that the potential to deploy gen AI at the edge and endpointhas only been apparent to developers for a matter of months. But when new generative AI use cases at the edge do emerge, one thing is almost certain: they will be running on an Ensemble device.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>When generative AI is performed in a data center, with access to practically unlimited compute and power resources, it is the modern miracle which provides A+ answers to examination questions, generates \u2018photographs\u2019 which look like images taken by a real camera, and predicts molecular behavior that previously would have needed to be observed in a laboratory experiment.<\/p>\n","protected":false},"author":1,"featured_media":1000002638,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"content-type":"","inline_featured_image":false,"footnotes":""},"categories":[42,102,107],"tags":[],"class_list":["post-1000002533","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-artificial-intelligence","category-ensemble","category-generative-ai"],"acf":[],"_links":{"self":[{"href":"https:\/\/alifsemi.com\/wp-json\/wp\/v2\/posts\/1000002533","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/alifsemi.com\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/alifsemi.com\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/alifsemi.com\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/alifsemi.com\/wp-json\/wp\/v2\/comments?post=1000002533"}],"version-history":[{"count":8,"href":"https:\/\/alifsemi.com\/wp-json\/wp\/v2\/posts\/1000002533\/revisions"}],"predecessor-version":[{"id":1000002640,"href":"https:\/\/alifsemi.com\/wp-json\/wp\/v2\/posts\/1000002533\/revisions\/1000002640"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/alifsemi.com\/wp-json\/wp\/v2\/media\/1000002638"}],"wp:attachment":[{"href":"https:\/\/alifsemi.com\/wp-json\/wp\/v2\/media?parent=1000002533"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/alifsemi.com\/wp-json\/wp\/v2\/categories?post=1000002533"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/alifsemi.com\/wp-json\/wp\/v2\/tags?post=1000002533"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}