{"id":168490,"date":"2023-06-06T16:00:00","date_gmt":"2023-06-06T15:00:00","guid":{"rendered":"https:\/\/liora.io\/en\/?p=168490"},"modified":"2026-02-06T09:01:52","modified_gmt":"2026-02-06T08:01:52","slug":"the-new-champion-of-open-source-llm-falcon","status":"publish","type":"post","link":"https:\/\/liora.io\/en\/the-new-champion-of-open-source-llm-falcon","title":{"rendered":"The new champion of open source LLM, Falcon"},"content":{"rendered":"<style>\n.elementor-heading-title{padding:0;margin:0;line-height:1}.elementor-widget-heading .elementor-heading-title[class*=elementor-size-]>a{color:inherit;font-size:inherit;line-height:inherit}.elementor-widget-heading .elementor-heading-title.elementor-size-small{font-size:15px}.elementor-widget-heading .elementor-heading-title.elementor-size-medium{font-size:19px}.elementor-widget-heading .elementor-heading-title.elementor-size-large{font-size:29px}.elementor-widget-heading .elementor-heading-title.elementor-size-xl{font-size:39px}.elementor-widget-heading .elementor-heading-title.elementor-size-xxl{font-size:59px}<\/style>\n<p><strong>In the open source community, LLaMA had the effect of a technological leap, giving independent developers access to a large GPT-level language model. Today, Abu Dhabi&#8217;s Institute of Innovation and Technology (IIT) unveils Falcon, an open source LLM that outperforms LLaMA.<\/strong><\/p>\n<h3>What is Falcon?<\/h3>\n<p>Falcon is presented as <a href=\"https:\/\/liora.io\/en\/large-language-models-llm-everything-you-need-to-know\" target=\"_blank\" rel=\"noopener\">the most powerful language model to date<\/a>, with three possible variants: <b>Falcon 1B, 7B and 40B<\/b>. Smaller than LLaMA, with <b>40 billion parameters<\/b> versus 65, it nevertheless outperforms the latter. According to <b>Hugging Face&#8217;s<\/b> evaluation criteria (IA2 Reasoning Challenge, HellaSwag, MMLU and TruthfulQA), Falcon 40B Instruct, a Falcon variant, and Falcon 40B are more powerful than LLaMA in terms of performance.<\/p>\n<p>This model is <b>multilingual<\/b>, understanding English, German, Spanish and French, and Dutch, Italian, Romanian, Portuguese, Czech, Polish and Swedish.<\/p>\n<p>To achieve this result, IIT used <b>a dataset of 1,000 billion tokens<\/b> and <b>a pipeline<\/b> capable of extracting verified content to ensure the quality of Falcon&#8217;s responses. This <a href=\"https:\/\/huggingface.co\/datasets\/tiiuae\/falcon-refinedweb\" target=\"_blank\" rel=\"noopener\">&#8220;refined-web&#8221;<\/a> dataset is also <b>open source<\/b>, so developers can train their IA to produce programs as powerful as, or even better than, those currently available.<\/p>\n<style>\n.elementor-widget-image{text-align:center}.elementor-widget-image a{display:inline-block}.elementor-widget-image a img[src$=\".svg\"]{width:48px}.elementor-widget-image img{vertical-align:middle;display:inline-block}<\/style>\n<p>\t\t\t\t\t\t\t\t\t\t\t\t<img decoding=\"async\" width=\"800\" height=\"800\" src=\"https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/4944103-1024x1024.jpg\" alt=\"\" loading=\"lazy\" srcset=\"https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/4944103-1024x1024.jpg 1024w, https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/4944103-300x300.jpg 300w, https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/4944103-150x150.jpg 150w, https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/4944103-768x768.jpg 768w, https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/4944103-1536x1536.jpg 1536w, https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/4944103.jpg 2000w\" sizes=\"(max-width: 800px) 100vw, 800px\"><\/p>\n<h3>How useful will it be?<\/h3>\n<p>Unlike its predecessor, developers will be able to use <a href=\"https:\/\/huggingface.co\/spaces\/HuggingFaceH4\/open_llm_leaderboard\" target=\"_blank\" rel=\"noopener\">Falcon<\/a> for<b> commercial purposes<\/b>. Although LLaMA is open source, these weights remain private for Meta, which limits its commercialization. This is why Falcon&#8217;s models, which use <b>a modified version of Apache 2.0<\/b>, can suit the user&#8217;s needs.<\/p>\n<p><img decoding=\"async\" width=\"800\" height=\"534\" src=\"https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/kaleidico-3V8xo5Gbusk-unsplash-1024x683.jpg\" alt=\"\" loading=\"lazy\" srcset=\"https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/kaleidico-3V8xo5Gbusk-unsplash-1024x683.jpg 1024w, https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/kaleidico-3V8xo5Gbusk-unsplash-300x200.jpg 300w, https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/kaleidico-3V8xo5Gbusk-unsplash-768x512.jpg 768w, https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/kaleidico-3V8xo5Gbusk-unsplash-1536x1024.jpg 1536w, https:\/\/liora.io\/app\/uploads\/sites\/9\/2023\/06\/kaleidico-3V8xo5Gbusk-unsplash.jpg 1920w\" sizes=\"(max-width: 800px) 100vw, 800px\"><\/p>\n<p><b>Developers trained<\/b> to design new artificial intelligence will then be able to use Falcon to create a generation of <b>even more powerful AIs<\/b>. That&#8217;s why, if you&#8217;ve enjoyed this article and are considering a career in Data Science, don&#8217;t hesitate to check out <a href=\"https:\/\/liora.io\/en\/blog-en\" target=\"_blank\" rel=\"noopener\">our articles<\/a> or <a href=\"\/en\/courses\/data-ai\/\" target=\"_blank\" rel=\"noopener\">our training offers<\/a> on Liora.<\/p>\n<p><i>Source : huggingface.co<\/i><\/p>\n","protected":false},"excerpt":{"rendered":"<p>In the open source community, LLaMA had the effect of a technological leap, giving independent developers access to a large GPT-level language model. Today, Abu Dhabi&#8217;s Institute of Innovation and Technology (IIT) unveils Falcon, an open source LLM that outperforms LLaMA. What is Falcon? Falcon is presented as the most powerful language model to date, [&hellip;]<\/p>\n","protected":false},"author":74,"featured_media":168491,"comment_status":"open","ping_status":"open","sticky":false,"template":"elementor_theme","format":"standard","meta":{"_acf_changed":false,"editor_notices":[],"footnotes":""},"categories":[2433],"class_list":["post-168490","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-data-ai"],"acf":[],"_links":{"self":[{"href":"https:\/\/liora.io\/en\/wp-json\/wp\/v2\/posts\/168490","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/liora.io\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/liora.io\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/liora.io\/en\/wp-json\/wp\/v2\/users\/74"}],"replies":[{"embeddable":true,"href":"https:\/\/liora.io\/en\/wp-json\/wp\/v2\/comments?post=168490"}],"version-history":[{"count":1,"href":"https:\/\/liora.io\/en\/wp-json\/wp\/v2\/posts\/168490\/revisions"}],"predecessor-version":[{"id":206398,"href":"https:\/\/liora.io\/en\/wp-json\/wp\/v2\/posts\/168490\/revisions\/206398"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/liora.io\/en\/wp-json\/wp\/v2\/media\/168491"}],"wp:attachment":[{"href":"https:\/\/liora.io\/en\/wp-json\/wp\/v2\/media?parent=168490"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/liora.io\/en\/wp-json\/wp\/v2\/categories?post=168490"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}