{"id":178675,"date":"2022-08-24T08:20:37","date_gmt":"2022-08-24T08:20:37","guid":{"rendered":"https:\/\/www.healthbenefitstimes.com\/glossary\/?p=178675"},"modified":"2022-12-06T07:22:16","modified_gmt":"2022-12-06T07:22:16","slug":"interrater-reliability","status":"publish","type":"post","link":"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/","title":{"rendered":"Interrater reliability"},"content":{"rendered":"<p>Consistency in the scoring of research data results by two or more data analysts. Many health care investigators analyze data that are graduated rather than binary. In an analysis of anxiety, for example, a graduated scale may rate research subjects as \u201cvery anxious,\u201d \u201csomewhat anxious,\u201d \u201cmildly anxious,\u201d or \u201cnot at all anxious,\u201d whereas a binary method of rating anxiety might include just the two categories \u201canxious\u201d and \u201cnot anxious.\u201d If the study is carried out and coded by more than one psychologist, the coders may not agree on the implementation of the graduated scale; some may interview a patient and find him or her \u201csomewhat\u201d anxious, while another might assess the patient as being \u2019\u2019very anxious.\u201d The congruence in the application of the rating scale by more than one psychologist constitutes its interrater reliability.<\/p>\n<hr \/>\n<p>The extent to which two independent parties, each using the same tool or examining the same data, arrive at matching conclusions. It is a measure of the agreement, consensus, or consistency of independent parties in using a common rating scale or instrument.<\/p>\n<hr \/>\n<p>&nbsp;<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Consistency in the scoring of research data results by two or more data analysts. Many health care investigators analyze data that are graduated rather than binary. In an analysis of anxiety, for example, a graduated scale may rate research subjects as \u201cvery anxious,\u201d \u201csomewhat anxious,\u201d \u201cmildly anxious,\u201d or \u201cnot at all anxious,\u201d whereas a binary [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[9],"tags":[],"class_list":["post-178675","post","type-post","status-publish","format-standard","hentry","category-i"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v21.1 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Interrater reliability - Definition of Interrater reliability<\/title>\n<meta name=\"description\" content=\"Consistency in the scoring of research data results by two or more data analysts. Many health care investigators analyze data that are graduated rather than binary. In an analysis of anxiety, for example, a graduated scale may rate research subjects as \u201cvery anxious,\u201d \u201csomewhat anxious,\u201d \u201cmildly anxious,\u201d or \u201cnot at all anxious,\u201d whereas a binary method of rating anxiety might include just the two categories \u201canxious\u201d and \u201cnot anxious.\u201d If the study is carried out and coded by more than one psychologist, the coders may not agree on the implementation of the graduated scale; some may interview a patient and find him or her \u201csomewhat\u201d anxious, while another might assess the patient as being \u2019\u2019very anxious.\u201d The congruence in the application of the rating scale by more than one psychologist constitutes its interrater reliability.The extent to which two independent parties, each using the same tool or examining the same data, arrive at matching conclusions. It is a measure of the agreement, consensus, or consistency of independent parties in using a common rating scale or instrument.\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Interrater reliability - Definition of Interrater reliability\" \/>\n<meta property=\"og:description\" content=\"Consistency in the scoring of research data results by two or more data analysts. Many health care investigators analyze data that are graduated rather than binary. In an analysis of anxiety, for example, a graduated scale may rate research subjects as \u201cvery anxious,\u201d \u201csomewhat anxious,\u201d \u201cmildly anxious,\u201d or \u201cnot at all anxious,\u201d whereas a binary method of rating anxiety might include just the two categories \u201canxious\u201d and \u201cnot anxious.\u201d If the study is carried out and coded by more than one psychologist, the coders may not agree on the implementation of the graduated scale; some may interview a patient and find him or her \u201csomewhat\u201d anxious, while another might assess the patient as being \u2019\u2019very anxious.\u201d The congruence in the application of the rating scale by more than one psychologist constitutes its interrater reliability.The extent to which two independent parties, each using the same tool or examining the same data, arrive at matching conclusions. It is a measure of the agreement, consensus, or consistency of independent parties in using a common rating scale or instrument.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/\" \/>\n<meta property=\"og:site_name\" content=\"Glossary\" \/>\n<meta property=\"article:published_time\" content=\"2022-08-24T08:20:37+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2022-12-06T07:22:16+00:00\" \/>\n<meta name=\"author\" content=\"Glossary\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Glossary\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"1 minute\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/\",\"url\":\"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/\",\"name\":\"Interrater reliability - Definition of Interrater reliability\",\"isPartOf\":{\"@id\":\"https:\/\/www.healthbenefitstimes.com\/glossary\/#website\"},\"datePublished\":\"2022-08-24T08:20:37+00:00\",\"dateModified\":\"2022-12-06T07:22:16+00:00\",\"author\":{\"@id\":\"https:\/\/www.healthbenefitstimes.com\/glossary\/#\/schema\/person\/ccfef987a4882e6356ae6d77d33e74c5\"},\"description\":\"Consistency in the scoring of research data results by two or more data analysts. Many health care investigators analyze data that are graduated rather than binary. In an analysis of anxiety, for example, a graduated scale may rate research subjects as \u201cvery anxious,\u201d \u201csomewhat anxious,\u201d \u201cmildly anxious,\u201d or \u201cnot at all anxious,\u201d whereas a binary method of rating anxiety might include just the two categories \u201canxious\u201d and \u201cnot anxious.\u201d If the study is carried out and coded by more than one psychologist, the coders may not agree on the implementation of the graduated scale; some may interview a patient and find him or her \u201csomewhat\u201d anxious, while another might assess the patient as being \u2019\u2019very anxious.\u201d The congruence in the application of the rating scale by more than one psychologist constitutes its interrater reliability.The extent to which two independent parties, each using the same tool or examining the same data, arrive at matching conclusions. It is a measure of the agreement, consensus, or consistency of independent parties in using a common rating scale or instrument.\",\"breadcrumb\":{\"@id\":\"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/www.healthbenefitstimes.com\/glossary\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Interrater reliability\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/www.healthbenefitstimes.com\/glossary\/#website\",\"url\":\"https:\/\/www.healthbenefitstimes.com\/glossary\/\",\"name\":\"Glossary\",\"description\":\"Difinitions\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/www.healthbenefitstimes.com\/glossary\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"en-US\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/www.healthbenefitstimes.com\/glossary\/#\/schema\/person\/ccfef987a4882e6356ae6d77d33e74c5\",\"name\":\"Glossary\",\"url\":\"https:\/\/www.healthbenefitstimes.com\/glossary\/author\/adminglossary\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Interrater reliability - Definition of Interrater reliability","description":"Consistency in the scoring of research data results by two or more data analysts. Many health care investigators analyze data that are graduated rather than binary. In an analysis of anxiety, for example, a graduated scale may rate research subjects as \u201cvery anxious,\u201d \u201csomewhat anxious,\u201d \u201cmildly anxious,\u201d or \u201cnot at all anxious,\u201d whereas a binary method of rating anxiety might include just the two categories \u201canxious\u201d and \u201cnot anxious.\u201d If the study is carried out and coded by more than one psychologist, the coders may not agree on the implementation of the graduated scale; some may interview a patient and find him or her \u201csomewhat\u201d anxious, while another might assess the patient as being \u2019\u2019very anxious.\u201d The congruence in the application of the rating scale by more than one psychologist constitutes its interrater reliability.The extent to which two independent parties, each using the same tool or examining the same data, arrive at matching conclusions. It is a measure of the agreement, consensus, or consistency of independent parties in using a common rating scale or instrument.","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/","og_locale":"en_US","og_type":"article","og_title":"Interrater reliability - Definition of Interrater reliability","og_description":"Consistency in the scoring of research data results by two or more data analysts. Many health care investigators analyze data that are graduated rather than binary. In an analysis of anxiety, for example, a graduated scale may rate research subjects as \u201cvery anxious,\u201d \u201csomewhat anxious,\u201d \u201cmildly anxious,\u201d or \u201cnot at all anxious,\u201d whereas a binary method of rating anxiety might include just the two categories \u201canxious\u201d and \u201cnot anxious.\u201d If the study is carried out and coded by more than one psychologist, the coders may not agree on the implementation of the graduated scale; some may interview a patient and find him or her \u201csomewhat\u201d anxious, while another might assess the patient as being \u2019\u2019very anxious.\u201d The congruence in the application of the rating scale by more than one psychologist constitutes its interrater reliability.The extent to which two independent parties, each using the same tool or examining the same data, arrive at matching conclusions. It is a measure of the agreement, consensus, or consistency of independent parties in using a common rating scale or instrument.","og_url":"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/","og_site_name":"Glossary","article_published_time":"2022-08-24T08:20:37+00:00","article_modified_time":"2022-12-06T07:22:16+00:00","author":"Glossary","twitter_card":"summary_large_image","twitter_misc":{"Written by":"Glossary","Est. reading time":"1 minute"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/","url":"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/","name":"Interrater reliability - Definition of Interrater reliability","isPartOf":{"@id":"https:\/\/www.healthbenefitstimes.com\/glossary\/#website"},"datePublished":"2022-08-24T08:20:37+00:00","dateModified":"2022-12-06T07:22:16+00:00","author":{"@id":"https:\/\/www.healthbenefitstimes.com\/glossary\/#\/schema\/person\/ccfef987a4882e6356ae6d77d33e74c5"},"description":"Consistency in the scoring of research data results by two or more data analysts. Many health care investigators analyze data that are graduated rather than binary. In an analysis of anxiety, for example, a graduated scale may rate research subjects as \u201cvery anxious,\u201d \u201csomewhat anxious,\u201d \u201cmildly anxious,\u201d or \u201cnot at all anxious,\u201d whereas a binary method of rating anxiety might include just the two categories \u201canxious\u201d and \u201cnot anxious.\u201d If the study is carried out and coded by more than one psychologist, the coders may not agree on the implementation of the graduated scale; some may interview a patient and find him or her \u201csomewhat\u201d anxious, while another might assess the patient as being \u2019\u2019very anxious.\u201d The congruence in the application of the rating scale by more than one psychologist constitutes its interrater reliability.The extent to which two independent parties, each using the same tool or examining the same data, arrive at matching conclusions. It is a measure of the agreement, consensus, or consistency of independent parties in using a common rating scale or instrument.","breadcrumb":{"@id":"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/www.healthbenefitstimes.com\/glossary\/interrater-reliability\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/www.healthbenefitstimes.com\/glossary\/"},{"@type":"ListItem","position":2,"name":"Interrater reliability"}]},{"@type":"WebSite","@id":"https:\/\/www.healthbenefitstimes.com\/glossary\/#website","url":"https:\/\/www.healthbenefitstimes.com\/glossary\/","name":"Glossary","description":"Difinitions","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/www.healthbenefitstimes.com\/glossary\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"en-US"},{"@type":"Person","@id":"https:\/\/www.healthbenefitstimes.com\/glossary\/#\/schema\/person\/ccfef987a4882e6356ae6d77d33e74c5","name":"Glossary","url":"https:\/\/www.healthbenefitstimes.com\/glossary\/author\/adminglossary\/"}]}},"_links":{"self":[{"href":"https:\/\/www.healthbenefitstimes.com\/glossary\/wp-json\/wp\/v2\/posts\/178675","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.healthbenefitstimes.com\/glossary\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.healthbenefitstimes.com\/glossary\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.healthbenefitstimes.com\/glossary\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.healthbenefitstimes.com\/glossary\/wp-json\/wp\/v2\/comments?post=178675"}],"version-history":[{"count":2,"href":"https:\/\/www.healthbenefitstimes.com\/glossary\/wp-json\/wp\/v2\/posts\/178675\/revisions"}],"predecessor-version":[{"id":197051,"href":"https:\/\/www.healthbenefitstimes.com\/glossary\/wp-json\/wp\/v2\/posts\/178675\/revisions\/197051"}],"wp:attachment":[{"href":"https:\/\/www.healthbenefitstimes.com\/glossary\/wp-json\/wp\/v2\/media?parent=178675"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.healthbenefitstimes.com\/glossary\/wp-json\/wp\/v2\/categories?post=178675"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.healthbenefitstimes.com\/glossary\/wp-json\/wp\/v2\/tags?post=178675"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}