{"id":10920,"date":"2024-02-01T13:58:38","date_gmt":"2024-02-01T13:58:38","guid":{"rendered":"https:\/\/cc.eurohpc.pl\/?page_id=10920"},"modified":"2024-02-01T14:08:39","modified_gmt":"2024-02-01T14:08:39","slug":"lepiszcze-en","status":"publish","type":"page","link":"https:\/\/cc.eurohpc.pl\/index.php\/en\/lepiszcze-en\/","title":{"rendered":"Lepiszcze- en"},"content":{"rendered":"\t\t<div data-elementor-type=\"wp-page\" data-elementor-id=\"10920\" class=\"elementor elementor-10920\" data-elementor-post-type=\"page\">\n\t\t\t\t\t\t<section class=\"elementor-section elementor-top-section elementor-element elementor-element-83598de elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"83598de\" data-element_type=\"section\" data-e-type=\"section\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-c7d28fa\" data-id=\"c7d28fa\" data-element_type=\"column\" data-e-type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-3e09220 elementor-widget elementor-widget-heading\" data-id=\"3e09220\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">How LEPISZCZE was Designed and Created: A Comprehensive Benchmark for Natural Language Processing Tasks in Polish<\/h2>\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<section class=\"elementor-section elementor-top-section elementor-element elementor-element-6ecff73 elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"6ecff73\" data-element_type=\"section\" data-e-type=\"section\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-c7099e0\" data-id=\"c7099e0\" data-element_type=\"column\" data-e-type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-0631b46 elementor-widget elementor-widget-text-editor\" data-id=\"0631b46\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t\t\t\t\t\t<p>The increasing availability of computational resources and training data for large language models raises the demand for robust evaluation environments to accurately assess progress in language modelling. In recent years, significant progress has been made in standardising evaluation environments for the English language, with environments like GLUE, SuperGLUE, and KILT becoming standard tools for assessing language models. In creating environments for other languages, many researchers have focused on replicating the GLUE environment, as exemplified by the Polish benchmark KLEJ.<br \/>The seminar will discuss the work dedicated to the LEPISZCZE tool. The authors provide an overview of efforts to create evaluation environments for low-resource languages, highlighting that many languages still lack a comprehensive set of test data to assess language models. They identify current gaps in evaluation environments and compare tasks available within these environments, referencing English and Chinese, languages with abundant training and testing resources.<br \/>The main result of the work is LEPISZCZE \u2013 a new evaluation environment for Polish language technology based on language modelling, featuring a diverse set of test tasks. The proposed environment is designed with flexibility to add tasks, introducing new language models, submit results, and versioning data and models. The authors, along with the environment, also present evaluations of several new language models, including both improved datasets from the existing literature and new test sets for novel tasks. The environment includes five existing datasets and eight new datasets that have not been used before to evaluate language models. The article also shares experiences and insights gained from developing the LEPISZCZE evaluation environment, serving as guidance for designers of similar environments in other languages with limited linguistic resources.<\/p>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-1683947 elementor-align-left elementor-widget elementor-widget-button\" data-id=\"1683947\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"button.default\">\n\t\t\t\t\t\t\t\t\t\t<a class=\"elementor-button elementor-button-link elementor-size-sm\" href=\"https:\/\/arxiv.org\/abs\/2211.13112\" target=\"_blank\">\n\t\t\t\t\t\t<span class=\"elementor-button-content-wrapper\">\n\t\t\t\t\t\t\t\t\t<span class=\"elementor-button-text\">READ MORE<\/span>\n\t\t\t\t\t<\/span>\n\t\t\t\t\t<\/a>\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<\/div>\n\t\t","protected":false},"excerpt":{"rendered":"<p>How LEPISZCZE was Designed and Created: A Comprehensive Benchmark for Natural Language Processing Tasks in Polish The increasing availability of computational resources and training data for large language models raises the demand for robust evaluation environments to accurately assess progress in language modelling. In recent years, significant progress has been made in standardising evaluation environments&hellip;&nbsp;<a href=\"https:\/\/cc.eurohpc.pl\/index.php\/en\/lepiszcze-en\/\" rel=\"bookmark\">Dowiedz si\u0119 wi\u0119cej &raquo;<span class=\"screen-reader-text\">Lepiszcze- en<\/span><\/a><\/p>\n","protected":false},"author":5,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"page-templates\/template-pagebuilder-full-width.php","meta":{"_monsterinsights_skip_tracking":false,"_monsterinsights_sitenote_active":false,"_monsterinsights_sitenote_note":"","_monsterinsights_sitenote_category":0,"neve_meta_sidebar":"","neve_meta_container":"","neve_meta_enable_content_width":"","neve_meta_content_width":0,"neve_meta_title_alignment":"","neve_meta_author_avatar":"","neve_post_elements_order":"","neve_meta_disable_header":"","neve_meta_disable_footer":"","neve_meta_disable_title":"","footnotes":""},"class_list":["post-10920","page","type-page","status-publish","hentry"],"_links":{"self":[{"href":"https:\/\/cc.eurohpc.pl\/index.php\/wp-json\/wp\/v2\/pages\/10920","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/cc.eurohpc.pl\/index.php\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/cc.eurohpc.pl\/index.php\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/cc.eurohpc.pl\/index.php\/wp-json\/wp\/v2\/users\/5"}],"replies":[{"embeddable":true,"href":"https:\/\/cc.eurohpc.pl\/index.php\/wp-json\/wp\/v2\/comments?post=10920"}],"version-history":[{"count":3,"href":"https:\/\/cc.eurohpc.pl\/index.php\/wp-json\/wp\/v2\/pages\/10920\/revisions"}],"predecessor-version":[{"id":10923,"href":"https:\/\/cc.eurohpc.pl\/index.php\/wp-json\/wp\/v2\/pages\/10920\/revisions\/10923"}],"wp:attachment":[{"href":"https:\/\/cc.eurohpc.pl\/index.php\/wp-json\/wp\/v2\/media?parent=10920"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}