<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE root>
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.1d1" xml:lang="en"><front><journal-meta><journal-id journal-id-type="publisher">REMEDIUM</journal-id><journal-title-group><journal-title>REMEDIUM</journal-title></journal-title-group><issn publication-format="print">1561-5936</issn><issn publication-format="electronic">2658-3534</issn><publisher><publisher-name>Joint-Stock Company Chicot</publisher-name></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">1728</article-id><article-id pub-id-type="doi">10.32687/1561-5936-2024-28-4-434-439</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Article</subject></subj-group></article-categories><title-group><article-title>Expert assessment of requirements for members of qualification commissions</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Кочубей</surname><given-names>Adelina V.</given-names></name><bio></bio><email>kochoubeya@gmail.com</email><xref ref-type="aff" rid="aff-1"/><xref ref-type="aff" rid="aff-2"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Мишарин</surname><given-names>Viktor M.</given-names></name><bio></bio><email>info@pulmonology-russia.ru</email><xref ref-type="aff" rid="aff-3"/><xref ref-type="aff" rid="aff-2"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Казаков</surname><given-names>Alexey S.</given-names></name><bio></bio><email>keyprojet@yandex.ru</email><xref ref-type="aff" rid="aff-4"/><xref ref-type="aff" rid="aff-2"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Кочубей</surname><given-names>Valentin V.</given-names></name><bio></bio><email>kochoubey@gmail.com</email><xref ref-type="aff" rid="aff-1"/><xref ref-type="aff" rid="aff-2"/></contrib></contrib-group><aff id="aff-1">Federal Scientific and Clinical Center of Specialized Types of Medical Care, Moscow, Russia</aff><aff id="aff-2">Russian University of Medicine, Moscow, Russia</aff><aff id="aff-3">Research Institute of Pulmonology, Moscow, Russia</aff><aff id="aff-4">Research Institute for Healthcare Organization and Medical Management, Moscow, Russia</aff><pub-date date-type="epub" iso-8601-date="2024-12-15" publication-format="electronic"><day>15</day><month>12</month><year>2024</year></pub-date><issue>4</issue><fpage>434</fpage><lpage>439</lpage><history><pub-date date-type="received" iso-8601-date="2025-10-15"><day>15</day><month>10</month><year>2025</year></pub-date></history><permissions><copyright-statement>Copyright © 2024,</copyright-statement><copyright-year>2024</copyright-year></permissions><abstract>The vulnerability of an assessment conducted by another person lies in the varying degree of rigor of the examiner, possibly his lack of competence and likely dependence on others. Ensuring objectivity is achieved by selecting examiners based on informal signs of professional development. Categorization implies an assessment of the qualifications of specialists directly by the members of the qualification commissions. The purpose of the study is to assess the ability of requirements for members of qualification commissions to ensure their unified expert level of qualification and independence. Methods. The expert survey was conducted by correspondence and consisted of one iteration. The expert assessment of the requirements for the members of the qualification commissions was carried out according to 3 parameters: The ability of the requirements to guarantee a single expert level of qualification of the commission members in professional activities (1) and related disciplines (2), as well as the independence of the commission members (3). A ten-point Staple scale was used. For the analysis, the sums of the estimates were calculated, the consistency of the estimates, the normality of the distribution and the differences in expert estimates for two or more criteria were checked. The threshold value of the error level, p, is less than 0.05. Results. Expert estimates are consistent, W = 0.877, p ≤ 0.001. The sum of the assessments of the ability of the requirements to ensure the expert level of qualification of a member of the commission in professional activity is equal to –193, in related disciplines — −199, independence — −299. Expert assessments of the ability of requirements to guarantee independence are lower than to guarantee an expert level in professional and related fields, U = 1372, p ≤ 0.001, for both pairs. The estimates of the first and second parameters do not differ significantly, U = 1964.5, p ≤ 0.915. The ability to guarantee expertise in professional and related fields is better assessed by accreditation requirements than by other requirements, p = 0.001. Conclusions. Most of the requirements imposed on members of qualification commissions do not guarantee their expert level in professional activities and related disciplines, as well as independence from third parties.</abstract><kwd-group xml:lang="en"><kwd>expert survey</kwd><kwd>categorization</kwd><kwd>qualification criteria</kwd><kwd>objectification of assessments</kwd></kwd-group><kwd-group xml:lang="ru"><kwd>экспертный опрос</kwd><kwd>аттестация</kwd><kwd>критерии квалификации</kwd><kwd>объективизация оценок</kwd></kwd-group></article-meta></front><body></body><back><ref-list><ref id="B1"><label>1.</label><mixed-citation>Gershuni O., Czabanowska K., Burazeri G. et al. Is there a golden recipe? A scoping review of public health workforce development. Eur. J. Public Health. 2019;29(3):401–408. DOI: 10.1093/eurpub/cky247</mixed-citation></ref><ref id="B2"><label>2.</label><mixed-citation>Staudenmann D., Waldner N., Lörwald A., Huwendiek S. Medical specialty certification exams studied according to the Ottawa Quality Criteria: a systematic review. BMC Med. Educ. 2023;23(1):619. DOI: 10.1186/s12909-023-04600-x</mixed-citation></ref><ref id="B3"><label>3.</label><mixed-citation>Zimina E. V., Kochubej A. V., Konanyhina A. K., Navarkin M. V. The domestic system of training and continuous professional development of medical workers: SWOT analysis. Modern problems of science and education. 2015;(4):445.</mixed-citation></ref><ref id="B4"><label>4.</label><mixed-citation>Thiessen N., Fischer M. R., Huwendiek S. Assessment methods in medical specialist assessments in the DACH region — overview, critical examination and recommendations for further development. GMS J. Med. Educ. 2019;36(6):78.</mixed-citation></ref><ref id="B5"><label>5.</label><mixed-citation>McManus I. C., Thompson M., Mollon J. Assessment of examiner leniency and stringency (&amp;apos;hawk-dove effect&amp;apos;) in the MRCP(UK) clinical examination (PACES) using multi-facet Rasch modelling. BMC Med. Educ. 2006;6:42. DOI: 10.1186/1472-6920-6-42</mixed-citation></ref><ref id="B6"><label>6.</label><mixed-citation>Fuchs J. W., Youmans Q. R. Mitigating bias in the era of virtual residency and fellowship interviews. J. Grad. Med. Edu. 2020;12:674–677.</mixed-citation></ref><ref id="B7"><label>7.</label><mixed-citation>Ryan T. Addressing bias and lack of objectivity in the Canadian resident matching process. CMAJ. 2018;190:E1211–E1212.</mixed-citation></ref><ref id="B8"><label>8.</label><mixed-citation>Osler W. Examinations, examiners and examinees. Lancet. 1913;136:1047–1050.</mixed-citation></ref><ref id="B9"><label>9.</label><mixed-citation>Fleming P. R., Manderson W. G., Matthews M. B. et al. Evolution of an examination, M. R. C. P. (UK). Br. Med. J. 1974;2(5910):99–107. DOI: 10.1136/bmj.2.5910.99</mixed-citation></ref><ref id="B10"><label>10.</label><mixed-citation>Wakeford R., Denney M., Ludka-Stempien K., Dacre J, McManus I. C. Cross-comparison of MRCGP &amp;amp; MRCP(UK) in a database linkage study of 2,284 candidates taking both examinations: assessment of validity and differential performance by ethnicity. BMC Med. Educ. 2015;15:1. DOI: 10.1186/s12909-014-0281-2</mixed-citation></ref><ref id="B11"><label>11.</label><mixed-citation>Finn Y., Cantillon P., Flaherty G. Exploration of a possible relationship between examiner stringency and personality factors in clinical assessments: a pilot study. BMC Med. Educ. 2014;14:1052. DOI: 10.1186/s12909-014-0280-3</mixed-citation></ref><ref id="B12"><label>12.</label><mixed-citation>Klein R., Snyder E. D., Koch J. et al. Analysis of narrative assessments of internal medicine resident performance: are there differences associated with gender or race and ethnicity? BMC Med. Educ. 2024;24(1):72. DOI: 10.1186/s12909-023-04970-2</mixed-citation></ref><ref id="B13"><label>13.</label><mixed-citation>Klein R., Snyder E. D., Koch J. et al. Exploring gender and thematic differences in qualitative assessments of internal medicine resident performance. BMC Med. Educ. 2023;23(1):932. DOI: 10.1186/s12909-023-04917-7</mixed-citation></ref><ref id="B14"><label>14.</label><mixed-citation>Locke R., Bell J., Scallan S. et al. The experience and professional development of medical appraisers. Educ Prim Care. 2018 Nov;29(6):351–356. DOI: 10.1080/14739879.2018.1514987</mixed-citation></ref><ref id="B15"><label>15.</label><mixed-citation>Atkinson A. R., Abbott C., Oswald A. et al. Strategies to enable transformation in medical education: faculty and trainee development in competence by design. Perspect. Med. Educ. 2024;13(1):85–94. DOI: 10.5334/pme.960</mixed-citation></ref><ref id="B16"><label>16.</label><mixed-citation>Schultz K. W., Kolomitro K., Koppula S., Bethune C. H. Competency-based faculty development: applying transformations from lessons learned in competency-based medical education. Can. Med. Educ. J. 2023;14(5):95–102. DOI: 10.36834/cmej.75768</mixed-citation></ref><ref id="B17"><label>17.</label><mixed-citation>Holmboe E. S., Ward D. S., Reznick R. K. et al. Faculty development in assessment: the missing link in competency-based medical education. Acad. Med. 2011;86(4):460–467. DOI: 10.1097/ACM.0b013e31820cb2a7</mixed-citation></ref><ref id="B18"><label>18.</label><mixed-citation>Yeates P., Sebok-Syer S. S. Hawks, doves and rasch decisions: understanding the influence of different cycles of an OSCE on students’ scores using many facet rasch modeling. Med. Teach. 2017;39:92–99.</mixed-citation></ref><ref id="B19"><label>19.</label><mixed-citation>Seaward J. R., Carter L. R., Nagarkar P., Zhang A. Y. Rating the rater: a technique for minimizing leniency bias in residency applications. Plast. Reconstr. Surg. Glob Open. 2023;11(4):e4892. DOI: 10.1097/GOX.0000000000004892</mixed-citation></ref><ref id="B20"><label>20.</label><mixed-citation>Kiraly L., Dewey E., Brasel K. Hawks and doves: adjusting for bias in residency interview scoring. J. Surg Educ. 2020;77:e132–e137.</mixed-citation></ref><ref id="B21"><label>21.</label><mixed-citation>Clement E. A., Oswald A., Ghosh S., Hamza D. M. Exploring the quality of feedback in entrustable professional activity narratives across 24 residency training programs. J Grad. Med. Educ. 2024;16(1):23–29. DOI: 10.4300/JGME-D-23-00210.1</mixed-citation></ref><ref id="B22"><label>22.</label><mixed-citation>Kornegay J. G., Kraut A., Manthey D. et al. Feedback in medical education: a critical appraisal. AEM Educ. Train. 2017;1:98–109. DOI: 10.1002/aet2.10024</mixed-citation></ref><ref id="B23"><label>23.</label><mixed-citation>Sheehan J. External examiners: roles and issues. J. Adv. Nurs. 1994;20(5):943–949. DOI: 10.1046/j.1365-2648.1994.20050943.x</mixed-citation></ref><ref id="B24"><label>24.</label><mixed-citation>Allen M., Russell T., Ford L. et al. Identification and evaluation of criterion measurement methods. Mil. Psychol. 2023;35(4):308–320. DOI: 10.1080/08995605.2022.2050165</mixed-citation></ref><ref id="B25"><label>25.</label><mixed-citation>Kuznetsova O. V., Samoilov A. S., Romanov S. V., Abaeva O. P. From certification to accreditation: the history of the development of domestic medical education and prospects for transition to the system continuing medical education. Extreme Medicine. 2018;20(4):551–558.</mixed-citation></ref><ref id="B26"><label>26.</label><mixed-citation>Lockyer J., Carraccio C., Chan M. K. et al. Core principles of assessment in competency-based medical education, Med. Teach. 2017;39(6):609–616. DOI: 10.1080/0142159X.2017.1315082</mixed-citation></ref><ref id="B27"><label>27.</label><mixed-citation>Wright C., Campbell J., McGowan L. et al. Interpreting multi-source feedback: online study of consensus and variation among GP appraisers. Br. J. Gen. Pract. 2016;66(645):e277–284. DOI: 10.3399/bjgp16X684373</mixed-citation></ref><ref id="B28"><label>28.</label><mixed-citation>Norcini J., Anderson M. B., Bollela V. et al. Consensus framework for good assessment. Med. Teach. 2018;40(11):1102–1109. DOI: 10.1080/0142159X.2018.1500016</mixed-citation></ref></ref-list></back></article>
