{"id":3875,"date":"2025-12-15T11:17:24","date_gmt":"2025-12-15T11:17:24","guid":{"rendered":"https:\/\/jita-au.com\/?p=3875"},"modified":"2025-12-15T13:48:22","modified_gmt":"2025-12-15T13:48:22","slug":"a-small-language-ai-model-in-the-bosnian-language","status":"publish","type":"post","link":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/","title":{"rendered":"A Small Language AI Model in the Bosnian Language"},"content":{"rendered":"\t\t<div data-elementor-type=\"wp-post\" data-elementor-id=\"3875\" class=\"elementor elementor-3875\" data-elementor-settings=\"{&quot;ha_cmc_init_switcher&quot;:&quot;no&quot;}\">\n\t\t\t\t\t\t<section class=\"elementor-section elementor-top-section elementor-element elementor-element-e232701 elementor-section-height-min-height elementor-section-full_width elementor-hidden-tablet elementor-hidden-mobile elementor-section-height-default elementor-section-items-middle wpr-particle-no wpr-jarallax-no wpr-parallax-no wpr-sticky-section-no\" data-id=\"e232701\" data-element_type=\"section\" data-settings=\"{&quot;background_background&quot;:&quot;classic&quot;,&quot;_ha_eqh_enable&quot;:false}\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-cb97594\" data-id=\"cb97594\" data-element_type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-a18c791 elementor-widget__width-inherit ha-has-bg-overlay elementor-widget elementor-widget-heading\" data-id=\"a18c791\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Vol. 15 No. 2 (2025): JITA - APEIRON<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-b2fff26 elementor-widget elementor-widget-heading\" data-id=\"b2fff26\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\"><b>Bo\u0161ko Jefi\u0107, Vlatko Bodul, Admir Agi\u0107<b><\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-52044bb elementor-widget-divider--view-line elementor-widget elementor-widget-divider\" data-id=\"52044bb\" data-element_type=\"widget\" data-widget_type=\"divider.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-divider\">\n\t\t\t<span class=\"elementor-divider-separator\">\n\t\t\t\t\t\t<\/span>\n\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-1c82606 elementor-widget elementor-widget-heading\" data-id=\"1c82606\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">A Small Language AI Model in the Bosnian Language<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-c297c9e elementor-widget-divider--view-line elementor-widget elementor-widget-divider\" data-id=\"c297c9e\" data-element_type=\"widget\" data-widget_type=\"divider.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-divider\">\n\t\t\t<span class=\"elementor-divider-separator\">\n\t\t\t\t\t\t<\/span>\n\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-b6ebd13 elementor-widget__width-inherit elementor-widget elementor-widget-heading\" data-id=\"b6ebd13\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Review paper<br>DOI: https:\/\/doi.org\/10.7251\/JIT2502128J<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-2f3b63f elementor-align-center elementor-widget elementor-widget-button\" data-id=\"2f3b63f\" data-element_type=\"widget\" data-widget_type=\"button.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<div class=\"elementor-button-wrapper\">\n\t\t\t\t\t<a class=\"elementor-button elementor-button-link elementor-size-sm\" href=\"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB-6.pdf\">\n\t\t\t\t\t\t<span class=\"elementor-button-content-wrapper\">\n\t\t\t\t\t\t\t\t\t<span class=\"elementor-button-text\">Download Article PDF<\/span>\n\t\t\t\t\t<\/span>\n\t\t\t\t\t<\/a>\n\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-cb7c4cb elementor-widget__width-inherit elementor-widget elementor-widget-heading\" data-id=\"cb7c4cb\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Abstract<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-61aebe3 elementor-widget__width-inherit elementor-widget elementor-widget-text-editor\" data-id=\"61aebe3\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<p>This study presents the development and evaluation of Mali Mujo, a small-scale language model optimized for the\nBosnian language, designed to operate efficiently on devices with limited computational resources. Leveraging the TinyLlama\narchitecture, the model demonstrates the feasibility of deploying natural language processing (NLP) applications in environments\nwith constrained memory and processing capabilities, specifically devices with 1 GB storage and 8 GB RAM. The system integrates\nLangchain agents and the DuckDuckGo API to enable real-time information retrieval, enhancing the model\u2019s responsiveness and\naccuracy in practical applications. The methodology involved training the TinyLlama model on a curated Bosnian dataset, followed\nby testing across diverse real-world scenarios in industry and administration. Performance metrics focused on accuracy, response\ntime, and computational efficiency, while additional evaluation considered user experience and adaptability to domain-specific\ntasks. The results indicate that Mali Mujo delivers rapid and reliable responses to user queries, with significant advantages in speed\nand resource efficiency compared to larger language models. The model effectively processes administrative requests, generates\ntechnical and market-related insights, and supports educational and governmental applications, highlighting its versatility. While\nsmall-scale models exhibit lower absolute accuracy than their larger counterparts, the study demonstrates that careful optimization\nand integration with external APIs can mitigate limitations, providing a balance between performance and accessibility.\nFurthermore, the model\u2019s design ensures user privacy and low energy consumption, contributing to sustainable and secure AI\ndeployment. Mali Mujo exemplifies the potential of small language models to enhance efficiency, accessibility, and usability in locallanguage\ncontexts. Its deployment provides a scalable, cost-effective solution for organizations with limited infrastructure, offering\nopportunities for further enhancement through expanded datasets, multilingual support, adaptive learning, and integration with\nemerging AI technologies. The findings underscore the practicality of small AI models in bridging the gap between advanced NLP\ncapabilities and resource-constrained environments.\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-5c5f19e elementor-widget elementor-widget-heading\" data-id=\"5c5f19e\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Keywords: Small language models, TinyLlama, Bosnian language, Langchain agents, Real-time information retrieval, AI in industry.<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-4df6678 elementor-widget elementor-widget-heading\" data-id=\"4df6678\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\"><b>Paper received:<\/b> 31.10.2025.<br><b>Paper accepted:<\/b> 27.11.2025.<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-6dd494a wpr-logo-position-center elementor-widget elementor-widget-wpr-logo\" data-id=\"6dd494a\" data-element_type=\"widget\" data-widget_type=\"wpr-logo.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\n\t\t\t<div class=\"wpr-logo elementor-clearfix\">\n\n\t\t\t\t\t\t\t\t<picture class=\"wpr-logo-image\">\n\t\t\t\t\t\n\t\t\t\t\t\t\t\t\t\t\n\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/jita-au.com\/wp-content\/uploads\/2024\/03\/cc-by-1.png\" alt=\"\">\n\n\t\t\t\t\t\t\t\t\t\t\t<a class=\"wpr-logo-url\" rel=\"home\" aria-label=\"\" href=\"https:\/\/jita-au.com\/\"><\/a>\n\t\t\t\t\t\t\t\t\t<\/picture>\n\t\t\t\t\n\t\t\t\t\n\t\t\t\t\t\t\t\t\t<a class=\"wpr-logo-url\" rel=\"home\" aria-label=\"\" href=\"https:\/\/jita-au.com\/\"><\/a>\n\t\t\t\t\n\t\t\t<\/div>\n\t\t\t\t\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-cb049d1 elementor-widget elementor-widget-shortcode\" data-id=\"cb049d1\" data-element_type=\"widget\" data-widget_type=\"shortcode.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-shortcode\"><div class=\"post-views content-post post-3875 entry-meta load-static\">\r\n\t\t\t\t<span class=\"post-views-icon dashicons dashicons-chart-bar\"><\/span> <span class=\"post-views-label\">Post Views:<\/span> <span class=\"post-views-count\">562<\/span>\r\n\t\t\t<\/div><\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-20a3f75 elementor-widget elementor-widget-shortcode\" data-id=\"20a3f75\" data-element_type=\"widget\" data-widget_type=\"shortcode.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-shortcode\">Downloaded Article PDF: <span class=\"snr-download-count-num\" data-post-id=\"0\">0<\/span> times\n<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<section class=\"elementor-section elementor-top-section elementor-element elementor-element-511d0df elementor-section-height-min-height elementor-section-full_width elementor-hidden-mobile elementor-hidden-desktop elementor-hidden-laptop elementor-section-height-default elementor-section-items-middle wpr-particle-no wpr-jarallax-no wpr-parallax-no wpr-sticky-section-no\" data-id=\"511d0df\" data-element_type=\"section\" data-settings=\"{&quot;background_background&quot;:&quot;classic&quot;,&quot;_ha_eqh_enable&quot;:false}\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-0acee9c\" data-id=\"0acee9c\" data-element_type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-c0e98b3 elementor-widget__width-inherit ha-has-bg-overlay elementor-widget elementor-widget-heading\" data-id=\"c0e98b3\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Vol. 15 No. 2 (2025): JITA - APEIRON<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-fce3cc9 elementor-widget elementor-widget-heading\" data-id=\"fce3cc9\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\"><b>Bo\u0161ko Jefi\u0107, Vlatko Bodul, Admir Agi\u0107<b><\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-68f1849 elementor-widget-divider--view-line elementor-widget elementor-widget-divider\" data-id=\"68f1849\" data-element_type=\"widget\" data-widget_type=\"divider.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-divider\">\n\t\t\t<span class=\"elementor-divider-separator\">\n\t\t\t\t\t\t<\/span>\n\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-552f766 elementor-widget elementor-widget-heading\" data-id=\"552f766\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">A Small Language AI Model in the Bosnian Language<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-7b2f501 elementor-widget-divider--view-line elementor-widget elementor-widget-divider\" data-id=\"7b2f501\" data-element_type=\"widget\" data-widget_type=\"divider.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-divider\">\n\t\t\t<span class=\"elementor-divider-separator\">\n\t\t\t\t\t\t<\/span>\n\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-5027126 elementor-widget__width-initial elementor-widget elementor-widget-heading\" data-id=\"5027126\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Review paper<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-4a5312c elementor-widget__width-initial elementor-widget elementor-widget-heading\" data-id=\"4a5312c\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">DOI: https:\/\/doi.org\/10.7251\/JIT2502128J<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-6b7f422 elementor-align-center elementor-widget elementor-widget-button\" data-id=\"6b7f422\" data-element_type=\"widget\" data-widget_type=\"button.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<div class=\"elementor-button-wrapper\">\n\t\t\t\t\t<a class=\"elementor-button elementor-button-link elementor-size-sm\" href=\"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB-6.pdf\">\n\t\t\t\t\t\t<span class=\"elementor-button-content-wrapper\">\n\t\t\t\t\t\t\t\t\t<span class=\"elementor-button-text\">Download Article PDF<\/span>\n\t\t\t\t\t<\/span>\n\t\t\t\t\t<\/a>\n\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-580ffd3 elementor-widget__width-inherit elementor-widget elementor-widget-heading\" data-id=\"580ffd3\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Abstract<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-dd3e258 elementor-widget__width-inherit elementor-widget elementor-widget-text-editor\" data-id=\"dd3e258\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<p>This study presents the development and evaluation of Mali Mujo, a small-scale language model optimized for the\nBosnian language, designed to operate efficiently on devices with limited computational resources. Leveraging the TinyLlama\narchitecture, the model demonstrates the feasibility of deploying natural language processing (NLP) applications in environments\nwith constrained memory and processing capabilities, specifically devices with 1 GB storage and 8 GB RAM. The system integrates\nLangchain agents and the DuckDuckGo API to enable real-time information retrieval, enhancing the model\u2019s responsiveness and\naccuracy in practical applications. The methodology involved training the TinyLlama model on a curated Bosnian dataset, followed\nby testing across diverse real-world scenarios in industry and administration. Performance metrics focused on accuracy, response\ntime, and computational efficiency, while additional evaluation considered user experience and adaptability to domain-specific\ntasks. The results indicate that Mali Mujo delivers rapid and reliable responses to user queries, with significant advantages in speed\nand resource efficiency compared to larger language models. The model effectively processes administrative requests, generates\ntechnical and market-related insights, and supports educational and governmental applications, highlighting its versatility. While\nsmall-scale models exhibit lower absolute accuracy than their larger counterparts, the study demonstrates that careful optimization\nand integration with external APIs can mitigate limitations, providing a balance between performance and accessibility.\nFurthermore, the model\u2019s design ensures user privacy and low energy consumption, contributing to sustainable and secure AI\ndeployment. Mali Mujo exemplifies the potential of small language models to enhance efficiency, accessibility, and usability in locallanguage\ncontexts. Its deployment provides a scalable, cost-effective solution for organizations with limited infrastructure, offering\nopportunities for further enhancement through expanded datasets, multilingual support, adaptive learning, and integration with\nemerging AI technologies. The findings underscore the practicality of small AI models in bridging the gap between advanced NLP\ncapabilities and resource-constrained environments.\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-62b72c3 elementor-widget elementor-widget-heading\" data-id=\"62b72c3\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Keywords: Small language models, TinyLlama, Bosnian language, Langchain agents, Real-time information retrieval, AI in industry.<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-a2636dd elementor-widget elementor-widget-heading\" data-id=\"a2636dd\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\"><b>Paper received:<\/b> 31.10.2025.<br><b>Paper accepted:<\/b> 27.11.2025.<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-f4dcd16 wpr-logo-position-center elementor-widget elementor-widget-wpr-logo\" data-id=\"f4dcd16\" data-element_type=\"widget\" data-widget_type=\"wpr-logo.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\n\t\t\t<div class=\"wpr-logo elementor-clearfix\">\n\n\t\t\t\t\t\t\t\t<picture class=\"wpr-logo-image\">\n\t\t\t\t\t\n\t\t\t\t\t\t\t\t\t\t\n\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/jita-au.com\/wp-content\/uploads\/2024\/03\/cc-by-1.png\" alt=\"\">\n\n\t\t\t\t\t\t\t\t\t\t\t<a class=\"wpr-logo-url\" rel=\"home\" aria-label=\"\" href=\"https:\/\/jita-au.com\/\"><\/a>\n\t\t\t\t\t\t\t\t\t<\/picture>\n\t\t\t\t\n\t\t\t\t\n\t\t\t\t\t\t\t\t\t<a class=\"wpr-logo-url\" rel=\"home\" aria-label=\"\" href=\"https:\/\/jita-au.com\/\"><\/a>\n\t\t\t\t\n\t\t\t<\/div>\n\t\t\t\t\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-0f73093 elementor-widget elementor-widget-shortcode\" data-id=\"0f73093\" data-element_type=\"widget\" data-widget_type=\"shortcode.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-shortcode\"><div class=\"post-views content-post post-3875 entry-meta load-static\">\r\n\t\t\t\t<span class=\"post-views-icon dashicons dashicons-chart-bar\"><\/span> <span class=\"post-views-label\">Post Views:<\/span> <span class=\"post-views-count\">562<\/span>\r\n\t\t\t<\/div><\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-975233a elementor-widget elementor-widget-shortcode\" data-id=\"975233a\" data-element_type=\"widget\" data-widget_type=\"shortcode.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-shortcode\">Downloaded Article PDF: <span class=\"snr-download-count-num\" data-post-id=\"0\">0<\/span> times\n<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<section class=\"elementor-section elementor-top-section elementor-element elementor-element-2d980b8 elementor-section-height-min-height elementor-section-full_width elementor-hidden-tablet elementor-hidden-desktop elementor-hidden-laptop elementor-section-height-default elementor-section-items-middle wpr-particle-no wpr-jarallax-no wpr-parallax-no wpr-sticky-section-no\" data-id=\"2d980b8\" data-element_type=\"section\" data-settings=\"{&quot;background_background&quot;:&quot;classic&quot;,&quot;_ha_eqh_enable&quot;:false}\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-d772afa\" data-id=\"d772afa\" data-element_type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-5597455 elementor-widget__width-inherit elementor-widget-mobile__width-inherit ha-has-bg-overlay elementor-widget elementor-widget-heading\" data-id=\"5597455\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Vol. 15 No. 2 (2025): JITA - APEIRON<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-2a5ea59 elementor-widget elementor-widget-heading\" data-id=\"2a5ea59\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\"><b>Bo\u0161ko Jefi\u0107, Vlatko Bodul, Admir Agi\u0107<b><\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-b0a72e2 elementor-widget-divider--view-line elementor-widget elementor-widget-divider\" data-id=\"b0a72e2\" data-element_type=\"widget\" data-widget_type=\"divider.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-divider\">\n\t\t\t<span class=\"elementor-divider-separator\">\n\t\t\t\t\t\t<\/span>\n\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-920c84d elementor-widget elementor-widget-heading\" data-id=\"920c84d\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">A Small Language AI Model in the Bosnian Language<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-96936f4 elementor-widget-divider--view-line elementor-widget elementor-widget-divider\" data-id=\"96936f4\" data-element_type=\"widget\" data-widget_type=\"divider.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-divider\">\n\t\t\t<span class=\"elementor-divider-separator\">\n\t\t\t\t\t\t<\/span>\n\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-f04febc elementor-widget__width-initial elementor-widget-mobile__width-inherit elementor-widget elementor-widget-heading\" data-id=\"f04febc\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Review paper<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-389228b elementor-widget__width-initial elementor-widget-mobile__width-inherit elementor-widget elementor-widget-heading\" data-id=\"389228b\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">DOI: https:\/\/doi.org\/10.7251\/JIT2502128J<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-ecaf2c2 elementor-align-center elementor-widget elementor-widget-button\" data-id=\"ecaf2c2\" data-element_type=\"widget\" data-widget_type=\"button.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<div class=\"elementor-button-wrapper\">\n\t\t\t\t\t<a class=\"elementor-button elementor-button-link elementor-size-sm\" href=\"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB-6.pdf\">\n\t\t\t\t\t\t<span class=\"elementor-button-content-wrapper\">\n\t\t\t\t\t\t\t\t\t<span class=\"elementor-button-text\">Download Article PDF<\/span>\n\t\t\t\t\t<\/span>\n\t\t\t\t\t<\/a>\n\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-80faf79 elementor-widget__width-inherit elementor-widget elementor-widget-heading\" data-id=\"80faf79\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Abstract<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-8cefffd elementor-widget__width-inherit elementor-widget elementor-widget-text-editor\" data-id=\"8cefffd\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<p>This study presents the development and evaluation of Mali Mujo, a small-scale language model optimized for the Bosnian language, designed to operate efficiently on devices with limited computational resources. Leveraging the TinyLlama architecture, the model demonstrates the feasibility of deploying natural language processing (NLP) applications in environments with constrained memory and processing capabilities, specifically devices with 1 GB storage and 8 GB RAM. The system integrates Langchain agents and the DuckDuckGo API to enable real-time information retrieval, enhancing the model\u2019s responsiveness and accuracy in practical applications. The methodology involved training the TinyLlama model on a curated Bosnian dataset, followed by testing across diverse real-world scenarios in industry and administration. Performance metrics focused on accuracy, response time, and computational efficiency, while additional evaluation considered user experience and adaptability to domain-specific tasks. The results indicate that Mali Mujo delivers rapid and reliable responses to user queries, with significant advantages in speed and resource efficiency compared to larger language models. The model effectively processes administrative requests, generates technical and market-related insights, and supports educational and governmental applications, highlighting its versatility. While small-scale models exhibit lower absolute accuracy than their larger counterparts, the study demonstrates that careful optimization and integration with external APIs can mitigate limitations, providing a balance between performance and accessibility. Furthermore, the model\u2019s design ensures user privacy and low energy consumption, contributing to sustainable and secure AI deployment. Mali Mujo exemplifies the potential of small language models to enhance efficiency, accessibility, and usability in locallanguage contexts. Its deployment provides a scalable, cost-effective solution for organizations with limited infrastructure, offering opportunities for further enhancement through expanded datasets, multilingual support, adaptive learning, and integration with emerging AI technologies. The findings underscore the practicality of small AI models in bridging the gap between advanced NLP capabilities and resource-constrained environments.<\/p>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-0c5e44b elementor-widget elementor-widget-heading\" data-id=\"0c5e44b\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Keywords: Small language models, TinyLlama, Bosnian language, Langchain agents, Real-time information retrieval, AI in industry.<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-8fcf84a elementor-widget elementor-widget-heading\" data-id=\"8fcf84a\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\"><b>Paper received:<\/b> 31.10.2025.<br><b>Paper accepted:<\/b> 27.11.2025.<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-dee52ab wpr-logo-position-center elementor-widget elementor-widget-wpr-logo\" data-id=\"dee52ab\" data-element_type=\"widget\" data-widget_type=\"wpr-logo.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\n\t\t\t<div class=\"wpr-logo elementor-clearfix\">\n\n\t\t\t\t\t\t\t\t<picture class=\"wpr-logo-image\">\n\t\t\t\t\t\n\t\t\t\t\t\t\t\t\t\t\n\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/jita-au.com\/wp-content\/uploads\/2024\/03\/cc-by-1.png\" alt=\"\">\n\n\t\t\t\t\t\t\t\t\t\t\t<a class=\"wpr-logo-url\" rel=\"home\" aria-label=\"\" href=\"https:\/\/jita-au.com\/\"><\/a>\n\t\t\t\t\t\t\t\t\t<\/picture>\n\t\t\t\t\n\t\t\t\t\n\t\t\t\t\t\t\t\t\t<a class=\"wpr-logo-url\" rel=\"home\" aria-label=\"\" href=\"https:\/\/jita-au.com\/\"><\/a>\n\t\t\t\t\n\t\t\t<\/div>\n\t\t\t\t\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-5cb9871 elementor-widget elementor-widget-shortcode\" data-id=\"5cb9871\" data-element_type=\"widget\" data-widget_type=\"shortcode.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-shortcode\"><div class=\"post-views content-post post-3875 entry-meta load-static\">\r\n\t\t\t\t<span class=\"post-views-icon dashicons dashicons-chart-bar\"><\/span> <span class=\"post-views-label\">Post Views:<\/span> <span class=\"post-views-count\">562<\/span>\r\n\t\t\t<\/div><\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-87a5ee0 elementor-widget elementor-widget-shortcode\" data-id=\"87a5ee0\" data-element_type=\"widget\" data-widget_type=\"shortcode.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-shortcode\">Downloaded Article PDF: <span class=\"snr-download-count-num\" data-post-id=\"0\">0<\/span> times\n<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<\/div>\n\t\t","protected":false},"excerpt":{"rendered":"<p>Vol. 15 No. 2 (2025): JITA &#8211; APEIRON Bo\u0161ko Jefi\u0107, Vlatko Bodul, Admir Agi\u0107 A Small Language AI Model in the Bosnian Language Review paperDOI: https:\/\/doi.org\/10.7251\/JIT2502128J Download Article PDF Abstract This study presents the development and evaluation of Mali Mujo, a small-scale language model optimized for the Bosnian language, designed to operate efficiently on devices with limited computational resources. Leveraging the TinyLlama architecture, the model demonstrates the feasibility of deploying natural language processing (NLP) applications in environments with constrained memory and processing capabilities, specifically devices with 1 GB storage and 8 GB RAM. The system integrates Langchain agents and the DuckDuckGo API to enable real-time information retrieval, enhancing the model\u2019s responsiveness and accuracy in practical applications. The methodology involved training the TinyLlama model on a curated Bosnian dataset, followed by testing across diverse real-world scenarios in industry and administration. Performance metrics focused on accuracy, response time, and computational efficiency, while additional evaluation considered user experience and adaptability to domain-specific tasks. The results indicate that Mali Mujo delivers rapid and reliable responses to user queries, with significant advantages in speed and resource efficiency compared to larger language models. The model effectively processes administrative requests, generates technical and market-related insights, and supports educational and governmental applications, highlighting its versatility. While small-scale models exhibit lower absolute accuracy than their larger counterparts, the study demonstrates that careful optimization and integration with external APIs can mitigate limitations, providing a balance between performance and accessibility. Furthermore, the model\u2019s design ensures user privacy and low energy consumption, contributing to sustainable and secure AI deployment. Mali Mujo exemplifies the potential of small language models to enhance efficiency, accessibility, and usability in locallanguage contexts. Its deployment provides a scalable, cost-effective solution for organizations with limited infrastructure, offering opportunities for further enhancement through expanded datasets, multilingual support, adaptive learning, and integration with emerging AI technologies. The findings underscore the practicality of small AI models in bridging the gap between advanced NLP capabilities and resource-constrained environments. Keywords: Small language models, TinyLlama, Bosnian language, Langchain agents, Real-time information retrieval, AI in industry. Paper received: 31.10.2025.Paper accepted: 27.11.2025. Vol. 15 No. 2 (2025): JITA &#8211; APEIRON Bo\u0161ko Jefi\u0107, Vlatko Bodul, Admir Agi\u0107 A Small Language AI Model in the Bosnian Language Review paper DOI: https:\/\/doi.org\/10.7251\/JIT2502128J Download Article PDF Abstract This study presents the development and evaluation of Mali Mujo, a small-scale language model optimized for the Bosnian language, designed to operate efficiently on devices with limited computational resources. Leveraging the TinyLlama architecture, the model demonstrates the feasibility of deploying natural language processing (NLP) applications in environments with constrained memory and processing capabilities, specifically devices with 1 GB storage and 8 GB RAM. The system integrates Langchain agents and the DuckDuckGo API to enable real-time information retrieval, enhancing the model\u2019s responsiveness and accuracy in practical applications. The methodology involved training the TinyLlama model on a curated Bosnian dataset, followed by testing across diverse real-world scenarios in industry and administration. Performance metrics focused on accuracy, response time, and computational efficiency, while additional evaluation considered user experience and adaptability to domain-specific tasks. The results indicate that Mali Mujo delivers rapid and reliable responses to user queries, with significant advantages in speed and resource efficiency compared to larger language models. The model effectively processes administrative requests, generates technical and market-related insights, and supports educational and governmental applications, highlighting its versatility. While small-scale models exhibit lower absolute accuracy than their larger counterparts, the study demonstrates that careful optimization and integration with external APIs can mitigate limitations, providing a balance between performance and accessibility. Furthermore, the model\u2019s design ensures user privacy and low energy consumption, contributing to sustainable and secure AI deployment. Mali Mujo exemplifies the potential of small language models to enhance efficiency, accessibility, and usability in locallanguage contexts. Its deployment provides a scalable, cost-effective solution for organizations with limited infrastructure, offering opportunities for further enhancement through expanded datasets, multilingual support, adaptive learning, and integration with emerging AI technologies. The findings underscore the practicality of small AI models in bridging the gap between advanced NLP capabilities and resource-constrained environments. Keywords: Small language models, TinyLlama, Bosnian language, Langchain agents, Real-time information retrieval, AI in industry. Paper received: 31.10.2025.Paper accepted: 27.11.2025. Vol. 15 No. 2 (2025): JITA &#8211; APEIRON Bo\u0161ko Jefi\u0107, Vlatko Bodul, Admir Agi\u0107 A Small Language AI Model in the Bosnian Language Review paper DOI: https:\/\/doi.org\/10.7251\/JIT2502128J Download Article PDF Abstract This study presents the development and evaluation of Mali Mujo, a small-scale language model optimized for the Bosnian language, designed to operate efficiently on devices with limited computational resources. Leveraging the TinyLlama architecture, the model demonstrates the feasibility of deploying natural language processing (NLP) applications in environments with constrained memory and processing capabilities, specifically devices with 1 GB storage and 8 GB RAM. The system integrates Langchain agents and the DuckDuckGo API to enable real-time information retrieval, enhancing the model\u2019s responsiveness and accuracy in practical applications. The methodology involved training the TinyLlama model on a curated Bosnian dataset, followed by testing across diverse real-world scenarios in industry and administration. Performance metrics focused on accuracy, response time, and computational efficiency, while additional evaluation considered user experience and adaptability to domain-specific tasks. The results indicate that Mali Mujo delivers rapid and reliable responses to user queries, with significant advantages in speed and resource efficiency compared to larger language models. The model effectively processes administrative requests, generates technical and market-related insights, and supports educational and governmental applications, highlighting its versatility. While small-scale models exhibit lower absolute accuracy than their larger counterparts, the study demonstrates that careful optimization and integration with external APIs can mitigate limitations, providing a balance between performance and accessibility. Furthermore, the model\u2019s design ensures user privacy and low energy consumption, contributing to sustainable and secure AI deployment. Mali Mujo exemplifies the potential of small language models to enhance efficiency, accessibility, and usability in locallanguage contexts. Its deployment provides a scalable, cost-effective solution for organizations with limited infrastructure, offering opportunities for further enhancement through expanded datasets, multilingual support, adaptive learning, and integration with emerging AI technologies. The findings<\/p>\n","protected":false},"author":1,"featured_media":3827,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"elementor_header_footer","format":"standard","meta":{"footnotes":""},"categories":[1],"tags":[],"class_list":["post-3875","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-uncategorized"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v26.8 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>A Small Language AI Model in the Bosnian Language - JITA -Journal of Information Technology and Application<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"A Small Language AI Model in the Bosnian Language - JITA -Journal of Information Technology and Application\" \/>\n<meta property=\"og:description\" content=\"Vol. 15 No. 2 (2025): JITA &#8211; APEIRON Bo\u0161ko Jefi\u0107, Vlatko Bodul, Admir Agi\u0107 A Small Language AI Model in the Bosnian Language Review paperDOI: https:\/\/doi.org\/10.7251\/JIT2502128J Download Article PDF Abstract This study presents the development and evaluation of Mali Mujo, a small-scale language model optimized for the Bosnian language, designed to operate efficiently on devices with limited computational resources. Leveraging the TinyLlama architecture, the model demonstrates the feasibility of deploying natural language processing (NLP) applications in environments with constrained memory and processing capabilities, specifically devices with 1 GB storage and 8 GB RAM. The system integrates Langchain agents and the DuckDuckGo API to enable real-time information retrieval, enhancing the model\u2019s responsiveness and accuracy in practical applications. The methodology involved training the TinyLlama model on a curated Bosnian dataset, followed by testing across diverse real-world scenarios in industry and administration. Performance metrics focused on accuracy, response time, and computational efficiency, while additional evaluation considered user experience and adaptability to domain-specific tasks. The results indicate that Mali Mujo delivers rapid and reliable responses to user queries, with significant advantages in speed and resource efficiency compared to larger language models. The model effectively processes administrative requests, generates technical and market-related insights, and supports educational and governmental applications, highlighting its versatility. While small-scale models exhibit lower absolute accuracy than their larger counterparts, the study demonstrates that careful optimization and integration with external APIs can mitigate limitations, providing a balance between performance and accessibility. Furthermore, the model\u2019s design ensures user privacy and low energy consumption, contributing to sustainable and secure AI deployment. Mali Mujo exemplifies the potential of small language models to enhance efficiency, accessibility, and usability in locallanguage contexts. Its deployment provides a scalable, cost-effective solution for organizations with limited infrastructure, offering opportunities for further enhancement through expanded datasets, multilingual support, adaptive learning, and integration with emerging AI technologies. The findings underscore the practicality of small AI models in bridging the gap between advanced NLP capabilities and resource-constrained environments. Keywords: Small language models, TinyLlama, Bosnian language, Langchain agents, Real-time information retrieval, AI in industry. Paper received: 31.10.2025.Paper accepted: 27.11.2025. Vol. 15 No. 2 (2025): JITA &#8211; APEIRON Bo\u0161ko Jefi\u0107, Vlatko Bodul, Admir Agi\u0107 A Small Language AI Model in the Bosnian Language Review paper DOI: https:\/\/doi.org\/10.7251\/JIT2502128J Download Article PDF Abstract This study presents the development and evaluation of Mali Mujo, a small-scale language model optimized for the Bosnian language, designed to operate efficiently on devices with limited computational resources. Leveraging the TinyLlama architecture, the model demonstrates the feasibility of deploying natural language processing (NLP) applications in environments with constrained memory and processing capabilities, specifically devices with 1 GB storage and 8 GB RAM. The system integrates Langchain agents and the DuckDuckGo API to enable real-time information retrieval, enhancing the model\u2019s responsiveness and accuracy in practical applications. The methodology involved training the TinyLlama model on a curated Bosnian dataset, followed by testing across diverse real-world scenarios in industry and administration. Performance metrics focused on accuracy, response time, and computational efficiency, while additional evaluation considered user experience and adaptability to domain-specific tasks. The results indicate that Mali Mujo delivers rapid and reliable responses to user queries, with significant advantages in speed and resource efficiency compared to larger language models. The model effectively processes administrative requests, generates technical and market-related insights, and supports educational and governmental applications, highlighting its versatility. While small-scale models exhibit lower absolute accuracy than their larger counterparts, the study demonstrates that careful optimization and integration with external APIs can mitigate limitations, providing a balance between performance and accessibility. Furthermore, the model\u2019s design ensures user privacy and low energy consumption, contributing to sustainable and secure AI deployment. Mali Mujo exemplifies the potential of small language models to enhance efficiency, accessibility, and usability in locallanguage contexts. Its deployment provides a scalable, cost-effective solution for organizations with limited infrastructure, offering opportunities for further enhancement through expanded datasets, multilingual support, adaptive learning, and integration with emerging AI technologies. The findings underscore the practicality of small AI models in bridging the gap between advanced NLP capabilities and resource-constrained environments. Keywords: Small language models, TinyLlama, Bosnian language, Langchain agents, Real-time information retrieval, AI in industry. Paper received: 31.10.2025.Paper accepted: 27.11.2025. Vol. 15 No. 2 (2025): JITA &#8211; APEIRON Bo\u0161ko Jefi\u0107, Vlatko Bodul, Admir Agi\u0107 A Small Language AI Model in the Bosnian Language Review paper DOI: https:\/\/doi.org\/10.7251\/JIT2502128J Download Article PDF Abstract This study presents the development and evaluation of Mali Mujo, a small-scale language model optimized for the Bosnian language, designed to operate efficiently on devices with limited computational resources. Leveraging the TinyLlama architecture, the model demonstrates the feasibility of deploying natural language processing (NLP) applications in environments with constrained memory and processing capabilities, specifically devices with 1 GB storage and 8 GB RAM. The system integrates Langchain agents and the DuckDuckGo API to enable real-time information retrieval, enhancing the model\u2019s responsiveness and accuracy in practical applications. The methodology involved training the TinyLlama model on a curated Bosnian dataset, followed by testing across diverse real-world scenarios in industry and administration. Performance metrics focused on accuracy, response time, and computational efficiency, while additional evaluation considered user experience and adaptability to domain-specific tasks. The results indicate that Mali Mujo delivers rapid and reliable responses to user queries, with significant advantages in speed and resource efficiency compared to larger language models. The model effectively processes administrative requests, generates technical and market-related insights, and supports educational and governmental applications, highlighting its versatility. While small-scale models exhibit lower absolute accuracy than their larger counterparts, the study demonstrates that careful optimization and integration with external APIs can mitigate limitations, providing a balance between performance and accessibility. Furthermore, the model\u2019s design ensures user privacy and low energy consumption, contributing to sustainable and secure AI deployment. Mali Mujo exemplifies the potential of small language models to enhance efficiency, accessibility, and usability in locallanguage contexts. Its deployment provides a scalable, cost-effective solution for organizations with limited infrastructure, offering opportunities for further enhancement through expanded datasets, multilingual support, adaptive learning, and integration with emerging AI technologies. The findings\" \/>\n<meta property=\"og:url\" content=\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/\" \/>\n<meta property=\"og:site_name\" content=\"JITA -Journal of Information Technology and Application\" \/>\n<meta property=\"article:published_time\" content=\"2025-12-15T11:17:24+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2025-12-15T13:48:22+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"612\" \/>\n\t<meta property=\"og:image:height\" content=\"805\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"admin\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"admin\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"5 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/\"},\"author\":{\"name\":\"admin\",\"@id\":\"https:\/\/jita-au.com\/#\/schema\/person\/d4becda53cfcbc99c449927eabf3877f\"},\"headline\":\"A Small Language AI Model in the Bosnian Language\",\"datePublished\":\"2025-12-15T11:17:24+00:00\",\"dateModified\":\"2025-12-15T13:48:22+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/\"},\"wordCount\":1076,\"publisher\":{\"@id\":\"https:\/\/jita-au.com\/#organization\"},\"image\":{\"@id\":\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB.jpg\",\"inLanguage\":\"en-US\"},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/\",\"url\":\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/\",\"name\":\"A Small Language AI Model in the Bosnian Language - JITA -Journal of Information Technology and Application\",\"isPartOf\":{\"@id\":\"https:\/\/jita-au.com\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB.jpg\",\"datePublished\":\"2025-12-15T11:17:24+00:00\",\"dateModified\":\"2025-12-15T13:48:22+00:00\",\"breadcrumb\":{\"@id\":\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#primaryimage\",\"url\":\"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB.jpg\",\"contentUrl\":\"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB.jpg\",\"width\":612,\"height\":805},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/jita-au.com\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"A Small Language AI Model in the Bosnian Language\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/jita-au.com\/#website\",\"url\":\"https:\/\/jita-au.com\/\",\"name\":\"JITA -Journal of Information Technology and Application\",\"description\":\"\",\"publisher\":{\"@id\":\"https:\/\/jita-au.com\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/jita-au.com\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/jita-au.com\/#organization\",\"name\":\"JITA -Journal of Information Technology and Application\",\"url\":\"https:\/\/jita-au.com\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/jita-au.com\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/jita-au.com\/wp-content\/uploads\/2024\/03\/cropped-JITA-logo-300px-1-1.jpg\",\"contentUrl\":\"https:\/\/jita-au.com\/wp-content\/uploads\/2024\/03\/cropped-JITA-logo-300px-1-1.jpg\",\"width\":300,\"height\":164,\"caption\":\"JITA -Journal of Information Technology and Application\"},\"image\":{\"@id\":\"https:\/\/jita-au.com\/#\/schema\/logo\/image\/\"}},{\"@type\":\"Person\",\"@id\":\"https:\/\/jita-au.com\/#\/schema\/person\/d4becda53cfcbc99c449927eabf3877f\",\"name\":\"admin\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/jita-au.com\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/fb1767673e75e9127846ff73b2b9e96214fba2d4675dc6799cec11e9b4380ca2?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/fb1767673e75e9127846ff73b2b9e96214fba2d4675dc6799cec11e9b4380ca2?s=96&d=mm&r=g\",\"caption\":\"admin\"},\"sameAs\":[\"https:\/\/jita-au.com\"],\"url\":\"https:\/\/jita-au.com\/index.php\/author\/jita-au-com\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"A Small Language AI Model in the Bosnian Language - JITA -Journal of Information Technology and Application","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/","og_locale":"en_US","og_type":"article","og_title":"A Small Language AI Model in the Bosnian Language - JITA -Journal of Information Technology and Application","og_description":"Vol. 15 No. 2 (2025): JITA &#8211; APEIRON Bo\u0161ko Jefi\u0107, Vlatko Bodul, Admir Agi\u0107 A Small Language AI Model in the Bosnian Language Review paperDOI: https:\/\/doi.org\/10.7251\/JIT2502128J Download Article PDF Abstract This study presents the development and evaluation of Mali Mujo, a small-scale language model optimized for the Bosnian language, designed to operate efficiently on devices with limited computational resources. Leveraging the TinyLlama architecture, the model demonstrates the feasibility of deploying natural language processing (NLP) applications in environments with constrained memory and processing capabilities, specifically devices with 1 GB storage and 8 GB RAM. The system integrates Langchain agents and the DuckDuckGo API to enable real-time information retrieval, enhancing the model\u2019s responsiveness and accuracy in practical applications. The methodology involved training the TinyLlama model on a curated Bosnian dataset, followed by testing across diverse real-world scenarios in industry and administration. Performance metrics focused on accuracy, response time, and computational efficiency, while additional evaluation considered user experience and adaptability to domain-specific tasks. The results indicate that Mali Mujo delivers rapid and reliable responses to user queries, with significant advantages in speed and resource efficiency compared to larger language models. The model effectively processes administrative requests, generates technical and market-related insights, and supports educational and governmental applications, highlighting its versatility. While small-scale models exhibit lower absolute accuracy than their larger counterparts, the study demonstrates that careful optimization and integration with external APIs can mitigate limitations, providing a balance between performance and accessibility. Furthermore, the model\u2019s design ensures user privacy and low energy consumption, contributing to sustainable and secure AI deployment. Mali Mujo exemplifies the potential of small language models to enhance efficiency, accessibility, and usability in locallanguage contexts. Its deployment provides a scalable, cost-effective solution for organizations with limited infrastructure, offering opportunities for further enhancement through expanded datasets, multilingual support, adaptive learning, and integration with emerging AI technologies. The findings underscore the practicality of small AI models in bridging the gap between advanced NLP capabilities and resource-constrained environments. Keywords: Small language models, TinyLlama, Bosnian language, Langchain agents, Real-time information retrieval, AI in industry. Paper received: 31.10.2025.Paper accepted: 27.11.2025. Vol. 15 No. 2 (2025): JITA &#8211; APEIRON Bo\u0161ko Jefi\u0107, Vlatko Bodul, Admir Agi\u0107 A Small Language AI Model in the Bosnian Language Review paper DOI: https:\/\/doi.org\/10.7251\/JIT2502128J Download Article PDF Abstract This study presents the development and evaluation of Mali Mujo, a small-scale language model optimized for the Bosnian language, designed to operate efficiently on devices with limited computational resources. Leveraging the TinyLlama architecture, the model demonstrates the feasibility of deploying natural language processing (NLP) applications in environments with constrained memory and processing capabilities, specifically devices with 1 GB storage and 8 GB RAM. The system integrates Langchain agents and the DuckDuckGo API to enable real-time information retrieval, enhancing the model\u2019s responsiveness and accuracy in practical applications. The methodology involved training the TinyLlama model on a curated Bosnian dataset, followed by testing across diverse real-world scenarios in industry and administration. Performance metrics focused on accuracy, response time, and computational efficiency, while additional evaluation considered user experience and adaptability to domain-specific tasks. The results indicate that Mali Mujo delivers rapid and reliable responses to user queries, with significant advantages in speed and resource efficiency compared to larger language models. The model effectively processes administrative requests, generates technical and market-related insights, and supports educational and governmental applications, highlighting its versatility. While small-scale models exhibit lower absolute accuracy than their larger counterparts, the study demonstrates that careful optimization and integration with external APIs can mitigate limitations, providing a balance between performance and accessibility. Furthermore, the model\u2019s design ensures user privacy and low energy consumption, contributing to sustainable and secure AI deployment. Mali Mujo exemplifies the potential of small language models to enhance efficiency, accessibility, and usability in locallanguage contexts. Its deployment provides a scalable, cost-effective solution for organizations with limited infrastructure, offering opportunities for further enhancement through expanded datasets, multilingual support, adaptive learning, and integration with emerging AI technologies. The findings underscore the practicality of small AI models in bridging the gap between advanced NLP capabilities and resource-constrained environments. Keywords: Small language models, TinyLlama, Bosnian language, Langchain agents, Real-time information retrieval, AI in industry. Paper received: 31.10.2025.Paper accepted: 27.11.2025. Vol. 15 No. 2 (2025): JITA &#8211; APEIRON Bo\u0161ko Jefi\u0107, Vlatko Bodul, Admir Agi\u0107 A Small Language AI Model in the Bosnian Language Review paper DOI: https:\/\/doi.org\/10.7251\/JIT2502128J Download Article PDF Abstract This study presents the development and evaluation of Mali Mujo, a small-scale language model optimized for the Bosnian language, designed to operate efficiently on devices with limited computational resources. Leveraging the TinyLlama architecture, the model demonstrates the feasibility of deploying natural language processing (NLP) applications in environments with constrained memory and processing capabilities, specifically devices with 1 GB storage and 8 GB RAM. The system integrates Langchain agents and the DuckDuckGo API to enable real-time information retrieval, enhancing the model\u2019s responsiveness and accuracy in practical applications. The methodology involved training the TinyLlama model on a curated Bosnian dataset, followed by testing across diverse real-world scenarios in industry and administration. Performance metrics focused on accuracy, response time, and computational efficiency, while additional evaluation considered user experience and adaptability to domain-specific tasks. The results indicate that Mali Mujo delivers rapid and reliable responses to user queries, with significant advantages in speed and resource efficiency compared to larger language models. The model effectively processes administrative requests, generates technical and market-related insights, and supports educational and governmental applications, highlighting its versatility. While small-scale models exhibit lower absolute accuracy than their larger counterparts, the study demonstrates that careful optimization and integration with external APIs can mitigate limitations, providing a balance between performance and accessibility. Furthermore, the model\u2019s design ensures user privacy and low energy consumption, contributing to sustainable and secure AI deployment. Mali Mujo exemplifies the potential of small language models to enhance efficiency, accessibility, and usability in locallanguage contexts. Its deployment provides a scalable, cost-effective solution for organizations with limited infrastructure, offering opportunities for further enhancement through expanded datasets, multilingual support, adaptive learning, and integration with emerging AI technologies. The findings","og_url":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/","og_site_name":"JITA -Journal of Information Technology and Application","article_published_time":"2025-12-15T11:17:24+00:00","article_modified_time":"2025-12-15T13:48:22+00:00","og_image":[{"width":612,"height":805,"url":"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB.jpg","type":"image\/jpeg"}],"author":"admin","twitter_card":"summary_large_image","twitter_misc":{"Written by":"admin","Est. reading time":"5 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#article","isPartOf":{"@id":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/"},"author":{"name":"admin","@id":"https:\/\/jita-au.com\/#\/schema\/person\/d4becda53cfcbc99c449927eabf3877f"},"headline":"A Small Language AI Model in the Bosnian Language","datePublished":"2025-12-15T11:17:24+00:00","dateModified":"2025-12-15T13:48:22+00:00","mainEntityOfPage":{"@id":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/"},"wordCount":1076,"publisher":{"@id":"https:\/\/jita-au.com\/#organization"},"image":{"@id":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#primaryimage"},"thumbnailUrl":"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB.jpg","inLanguage":"en-US"},{"@type":"WebPage","@id":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/","url":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/","name":"A Small Language AI Model in the Bosnian Language - JITA -Journal of Information Technology and Application","isPartOf":{"@id":"https:\/\/jita-au.com\/#website"},"primaryImageOfPage":{"@id":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#primaryimage"},"image":{"@id":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#primaryimage"},"thumbnailUrl":"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB.jpg","datePublished":"2025-12-15T11:17:24+00:00","dateModified":"2025-12-15T13:48:22+00:00","breadcrumb":{"@id":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#primaryimage","url":"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB.jpg","contentUrl":"https:\/\/jita-au.com\/wp-content\/uploads\/2025\/12\/Pages-from-JITA_Vol-15_Issue-2-WEB.jpg","width":612,"height":805},{"@type":"BreadcrumbList","@id":"https:\/\/jita-au.com\/index.php\/2025\/12\/15\/a-small-language-ai-model-in-the-bosnian-language\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/jita-au.com\/"},{"@type":"ListItem","position":2,"name":"A Small Language AI Model in the Bosnian Language"}]},{"@type":"WebSite","@id":"https:\/\/jita-au.com\/#website","url":"https:\/\/jita-au.com\/","name":"JITA -Journal of Information Technology and Application","description":"","publisher":{"@id":"https:\/\/jita-au.com\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/jita-au.com\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/jita-au.com\/#organization","name":"JITA -Journal of Information Technology and Application","url":"https:\/\/jita-au.com\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/jita-au.com\/#\/schema\/logo\/image\/","url":"https:\/\/jita-au.com\/wp-content\/uploads\/2024\/03\/cropped-JITA-logo-300px-1-1.jpg","contentUrl":"https:\/\/jita-au.com\/wp-content\/uploads\/2024\/03\/cropped-JITA-logo-300px-1-1.jpg","width":300,"height":164,"caption":"JITA -Journal of Information Technology and Application"},"image":{"@id":"https:\/\/jita-au.com\/#\/schema\/logo\/image\/"}},{"@type":"Person","@id":"https:\/\/jita-au.com\/#\/schema\/person\/d4becda53cfcbc99c449927eabf3877f","name":"admin","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/jita-au.com\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/fb1767673e75e9127846ff73b2b9e96214fba2d4675dc6799cec11e9b4380ca2?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/fb1767673e75e9127846ff73b2b9e96214fba2d4675dc6799cec11e9b4380ca2?s=96&d=mm&r=g","caption":"admin"},"sameAs":["https:\/\/jita-au.com"],"url":"https:\/\/jita-au.com\/index.php\/author\/jita-au-com\/"}]}},"_links":{"self":[{"href":"https:\/\/jita-au.com\/index.php\/wp-json\/wp\/v2\/posts\/3875","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/jita-au.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/jita-au.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/jita-au.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/jita-au.com\/index.php\/wp-json\/wp\/v2\/comments?post=3875"}],"version-history":[{"count":7,"href":"https:\/\/jita-au.com\/index.php\/wp-json\/wp\/v2\/posts\/3875\/revisions"}],"predecessor-version":[{"id":3954,"href":"https:\/\/jita-au.com\/index.php\/wp-json\/wp\/v2\/posts\/3875\/revisions\/3954"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/jita-au.com\/index.php\/wp-json\/wp\/v2\/media\/3827"}],"wp:attachment":[{"href":"https:\/\/jita-au.com\/index.php\/wp-json\/wp\/v2\/media?parent=3875"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/jita-au.com\/index.php\/wp-json\/wp\/v2\/categories?post=3875"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/jita-au.com\/index.php\/wp-json\/wp\/v2\/tags?post=3875"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}