{"id":15682,"date":"2026-01-07T08:31:24","date_gmt":"2026-01-07T07:31:24","guid":{"rendered":"https:\/\/agcommtech.de\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/"},"modified":"2026-01-07T08:31:36","modified_gmt":"2026-01-07T07:31:36","slug":"survey-and-analysis-of-hallucinations-in-large-language-models","status":"publish","type":"post","link":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/","title":{"rendered":"Survey and analysis of hallucinations in large language models"},"content":{"rendered":"\n<h2 class=\"wp-block-heading\"><strong><strong><strong>Survey and analysis of hallucinations in large language models &#8211; Study by Anh-Hoang, Tran &amp; Nguyen<\/strong><\/strong><\/strong><\/h2>\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"1626\" height=\"994\" src=\"https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025.png\" alt=\"\" class=\"wp-image-15679\" srcset=\"https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025.png 1626w, https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025-300x183.png 300w, https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025-1024x626.png 1024w, https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025-768x469.png 768w, https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025-1536x939.png 1536w, https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025-900x550.png 900w\" sizes=\"auto, (max-width: 1626px) 100vw, 1626px\" \/><\/figure>\n<\/div>\n<div style=\"height:25px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n<p>The study by Anh-Hoang, Tran and Nguyen, published in <em>Frontiers in Artificial Intelligence<\/em> in 2025, analyzes the problem of <strong>hallucinations in large language models<\/strong> (LLMs), i.e. false or unfounded statements presented as fact by AI systems. The aim of the work is to understand the extent to which such errors are influenced by the design of prompts and where the limits of prompting lie. <\/p>\n\n<p>A key finding of the study is that the wording of prompts has a significant influence on the reliability of AI-generated responses. Unclear or ambiguous instructions significantly increase the probability of hallucinations. In contrast, structured, explicit and context-rich prompts lead to measurably better results. The use of <strong>chain-of-thought prompting<\/strong>, in which the AI is guided to think step by step, is particularly emphasized. This technique significantly reduces hallucinations in many models.    <\/p>\n\n<p>At the same time, the study clearly shows that prompting alone is not a panacea. With certain models, hallucinations occur even when the prompts are precise and well-structured. In these cases, the causes lie in factors internal to the model, such as training data or limited factual knowledge. The authors therefore emphasize that hallucinations can always have both prompt-dependent and model-related causes.   <\/p>\n\n<p>Overall, the study concludes that a reduction in hallucinations can only be achieved through the interaction of several measures. Prompt optimization is an effective but limited instrument that needs to be supplemented by model-side improvements and additional control mechanisms. <\/p>\n\n<div style=\"height:30px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n<div class=\"wp-block-buttons is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button has-custom-width wp-block-button__width-100\"><a class=\"wp-block-button__link has-background wp-element-button\" href=\"https:\/\/www.frontiersin.org\/journals\/artificial-intelligence\/articles\/10.3389\/frai.2025.1622292\/full\" style=\"border-radius:0px;background-color:#274856\" target=\"_blank\" rel=\"noreferrer noopener\"><strong>TO THE READING TIP<\/strong><\/a><\/div>\n<\/div>\n\n<div style=\"height:30px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n<div class=\"wp-block-buttons is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button has-custom-width wp-block-button__width-100 is-style-outline is-style-outline--1\"><a class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/agcommtech.de\/en\/reading-tips-from-the-commtech-wg\/\" style=\"border-radius:0px\"><strong>ALL READING TIPS<\/strong><\/a><\/div>\n<\/div>\n\n<p><\/p>\n","protected":false},"excerpt":{"rendered":"<p>The study by Anh-Hoang, Tran and Nguyen, published in Frontiers in Artificial Intelligence in 2025, analyzes the problem of hallucinations in large language models (LLMs), i.e. false or unfounded statements presented as fact by AI systems. The aim of the work is to understand the extent to which such errors are influenced by the design of prompts and where the limits of prompting lie. <\/p>\n","protected":false},"author":2,"featured_media":15680,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[96],"tags":[],"ppma_author":[123],"class_list":["post-15682","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-reading-tips"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.3 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Survey and analysis of hallucinations in large language models - AG CommTech<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Survey and analysis of hallucinations in large language models - AG CommTech\" \/>\n<meta property=\"og:description\" content=\"The study by Anh-Hoang, Tran and Nguyen, published in Frontiers in Artificial Intelligence in 2025, analyzes the problem of hallucinations in large language models (LLMs), i.e. false or unfounded statements presented as fact by AI systems. The aim of the work is to understand the extent to which such errors are influenced by the design of prompts and where the limits of prompting lie.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/\" \/>\n<meta property=\"og:site_name\" content=\"AG CommTech\" \/>\n<meta property=\"article:published_time\" content=\"2026-01-07T07:31:24+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2026-01-07T07:31:36+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025.png\" \/>\n\t<meta property=\"og:image:width\" content=\"1626\" \/>\n\t<meta property=\"og:image:height\" content=\"994\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"Die Redaktion\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Die Redaktion\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"1 minute\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/\"},\"author\":{\"name\":\"Die Redaktion\",\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/#\\\/schema\\\/person\\\/62f8f4c6953ad8fa88a2dfd6d9d4ece1\"},\"headline\":\"Survey and analysis of hallucinations in large language models\",\"datePublished\":\"2026-01-07T07:31:24+00:00\",\"dateModified\":\"2026-01-07T07:31:36+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/\"},\"wordCount\":268,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/#organization\"},\"image\":{\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/agcommtech.de\\\/wp-content\\\/uploads\\\/2026\\\/01\\\/lesetipp-2-januar-2025.png\",\"articleSection\":[\"READING TIPS\"],\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/\",\"url\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/\",\"name\":\"Survey and analysis of hallucinations in large language models - AG CommTech\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/agcommtech.de\\\/wp-content\\\/uploads\\\/2026\\\/01\\\/lesetipp-2-januar-2025.png\",\"datePublished\":\"2026-01-07T07:31:24+00:00\",\"dateModified\":\"2026-01-07T07:31:36+00:00\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/#primaryimage\",\"url\":\"https:\\\/\\\/agcommtech.de\\\/wp-content\\\/uploads\\\/2026\\\/01\\\/lesetipp-2-januar-2025.png\",\"contentUrl\":\"https:\\\/\\\/agcommtech.de\\\/wp-content\\\/uploads\\\/2026\\\/01\\\/lesetipp-2-januar-2025.png\",\"width\":1626,\"height\":994},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/2026\\\/01\\\/07\\\/survey-and-analysis-of-hallucinations-in-large-language-models\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Startseite\",\"item\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/ag-commtech\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Survey and analysis of hallucinations in large language models\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/#website\",\"url\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/\",\"name\":\"AG CommTech\",\"description\":\"\",\"publisher\":{\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/#organization\",\"name\":\"AG CommTech\",\"url\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/agcommtech.de\\\/wp-content\\\/uploads\\\/2023\\\/05\\\/IMWF_220617_CommTech-Logo.png\",\"contentUrl\":\"https:\\\/\\\/agcommtech.de\\\/wp-content\\\/uploads\\\/2023\\\/05\\\/IMWF_220617_CommTech-Logo.png\",\"width\":1795,\"height\":691,\"caption\":\"AG CommTech\"},\"image\":{\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/#\\\/schema\\\/logo\\\/image\\\/\"}},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/#\\\/schema\\\/person\\\/62f8f4c6953ad8fa88a2dfd6d9d4ece1\",\"name\":\"Die Redaktion\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/203782fd9fd2ba9f595273ce30bd34f2a74a3559f20866d98a43495f0ac67353?s=96&d=mm&r=g704cf4e8ed0a9901b4a2fc54aebbe336\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/203782fd9fd2ba9f595273ce30bd34f2a74a3559f20866d98a43495f0ac67353?s=96&d=mm&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/203782fd9fd2ba9f595273ce30bd34f2a74a3559f20866d98a43495f0ac67353?s=96&d=mm&r=g\",\"caption\":\"Die Redaktion\"},\"url\":\"https:\\\/\\\/agcommtech.de\\\/en\\\/author\\\/die-redaktion\\\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Survey and analysis of hallucinations in large language models - AG CommTech","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/","og_locale":"en_US","og_type":"article","og_title":"Survey and analysis of hallucinations in large language models - AG CommTech","og_description":"The study by Anh-Hoang, Tran and Nguyen, published in Frontiers in Artificial Intelligence in 2025, analyzes the problem of hallucinations in large language models (LLMs), i.e. false or unfounded statements presented as fact by AI systems. The aim of the work is to understand the extent to which such errors are influenced by the design of prompts and where the limits of prompting lie.","og_url":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/","og_site_name":"AG CommTech","article_published_time":"2026-01-07T07:31:24+00:00","article_modified_time":"2026-01-07T07:31:36+00:00","og_image":[{"width":1626,"height":994,"url":"https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025.png","type":"image\/png"}],"author":"Die Redaktion","twitter_card":"summary_large_image","twitter_misc":{"Written by":"Die Redaktion","Est. reading time":"1 minute"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/#article","isPartOf":{"@id":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/"},"author":{"name":"Die Redaktion","@id":"https:\/\/agcommtech.de\/en\/#\/schema\/person\/62f8f4c6953ad8fa88a2dfd6d9d4ece1"},"headline":"Survey and analysis of hallucinations in large language models","datePublished":"2026-01-07T07:31:24+00:00","dateModified":"2026-01-07T07:31:36+00:00","mainEntityOfPage":{"@id":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/"},"wordCount":268,"commentCount":0,"publisher":{"@id":"https:\/\/agcommtech.de\/en\/#organization"},"image":{"@id":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/#primaryimage"},"thumbnailUrl":"https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025.png","articleSection":["READING TIPS"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/","url":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/","name":"Survey and analysis of hallucinations in large language models - AG CommTech","isPartOf":{"@id":"https:\/\/agcommtech.de\/en\/#website"},"primaryImageOfPage":{"@id":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/#primaryimage"},"image":{"@id":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/#primaryimage"},"thumbnailUrl":"https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025.png","datePublished":"2026-01-07T07:31:24+00:00","dateModified":"2026-01-07T07:31:36+00:00","breadcrumb":{"@id":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/#primaryimage","url":"https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025.png","contentUrl":"https:\/\/agcommtech.de\/wp-content\/uploads\/2026\/01\/lesetipp-2-januar-2025.png","width":1626,"height":994},{"@type":"BreadcrumbList","@id":"https:\/\/agcommtech.de\/en\/2026\/01\/07\/survey-and-analysis-of-hallucinations-in-large-language-models\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Startseite","item":"https:\/\/agcommtech.de\/en\/ag-commtech\/"},{"@type":"ListItem","position":2,"name":"Survey and analysis of hallucinations in large language models"}]},{"@type":"WebSite","@id":"https:\/\/agcommtech.de\/en\/#website","url":"https:\/\/agcommtech.de\/en\/","name":"AG CommTech","description":"","publisher":{"@id":"https:\/\/agcommtech.de\/en\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/agcommtech.de\/en\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/agcommtech.de\/en\/#organization","name":"AG CommTech","url":"https:\/\/agcommtech.de\/en\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/agcommtech.de\/en\/#\/schema\/logo\/image\/","url":"https:\/\/agcommtech.de\/wp-content\/uploads\/2023\/05\/IMWF_220617_CommTech-Logo.png","contentUrl":"https:\/\/agcommtech.de\/wp-content\/uploads\/2023\/05\/IMWF_220617_CommTech-Logo.png","width":1795,"height":691,"caption":"AG CommTech"},"image":{"@id":"https:\/\/agcommtech.de\/en\/#\/schema\/logo\/image\/"}},{"@type":"Person","@id":"https:\/\/agcommtech.de\/en\/#\/schema\/person\/62f8f4c6953ad8fa88a2dfd6d9d4ece1","name":"Die Redaktion","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/secure.gravatar.com\/avatar\/203782fd9fd2ba9f595273ce30bd34f2a74a3559f20866d98a43495f0ac67353?s=96&d=mm&r=g704cf4e8ed0a9901b4a2fc54aebbe336","url":"https:\/\/secure.gravatar.com\/avatar\/203782fd9fd2ba9f595273ce30bd34f2a74a3559f20866d98a43495f0ac67353?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/203782fd9fd2ba9f595273ce30bd34f2a74a3559f20866d98a43495f0ac67353?s=96&d=mm&r=g","caption":"Die Redaktion"},"url":"https:\/\/agcommtech.de\/en\/author\/die-redaktion\/"}]}},"authors":[{"term_id":123,"user_id":2,"is_guest":0,"slug":"die-redaktion","display_name":"Die Redaktion","avatar_url":"https:\/\/secure.gravatar.com\/avatar\/203782fd9fd2ba9f595273ce30bd34f2a74a3559f20866d98a43495f0ac67353?s=96&d=mm&r=g","0":null,"1":"","2":"","3":"","4":"","5":"","6":"","7":"","8":""}],"_links":{"self":[{"href":"https:\/\/agcommtech.de\/en\/wp-json\/wp\/v2\/posts\/15682","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/agcommtech.de\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/agcommtech.de\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/agcommtech.de\/en\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/agcommtech.de\/en\/wp-json\/wp\/v2\/comments?post=15682"}],"version-history":[{"count":1,"href":"https:\/\/agcommtech.de\/en\/wp-json\/wp\/v2\/posts\/15682\/revisions"}],"predecessor-version":[{"id":15683,"href":"https:\/\/agcommtech.de\/en\/wp-json\/wp\/v2\/posts\/15682\/revisions\/15683"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/agcommtech.de\/en\/wp-json\/wp\/v2\/media\/15680"}],"wp:attachment":[{"href":"https:\/\/agcommtech.de\/en\/wp-json\/wp\/v2\/media?parent=15682"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/agcommtech.de\/en\/wp-json\/wp\/v2\/categories?post=15682"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/agcommtech.de\/en\/wp-json\/wp\/v2\/tags?post=15682"},{"taxonomy":"author","embeddable":true,"href":"https:\/\/agcommtech.de\/en\/wp-json\/wp\/v2\/ppma_author?post=15682"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}