How can we help you?
Click on the section to view help questions and answers:
Periodically checks the accuracy of the results
First Steps to Implementation:
The administrator is responsible for managing the relationship with ControlLab, keeping the registration data up to date, managing and delegating tasks to the Online System, receiving and distributing materials and correspondence, ensuring deadlines and analyzing results.
One of the main tasks is to promote the dissemination of information, starting with the purpose, operating rules and evaluation criteria of the PT. For this, it is important to ensure that all laboratory stakeholders read and understand the Gibi Quality Control information, the Participant Manual and the “Questions and Answers” available on the site.
To manage the day-to-day of the program there are the warning emails and the homepage of the Online System, which describes the rounds in progress with the deadline for reporting results and recently published evaluations.
For the monitoring the evaluations, there is the management report and the general accumulation. It is important to conduct a critical analysis of the results together with those involved and analyze the causes for inappropriate results immediately, based on the evaluation report, results profile and internal documents/records. To help you in this step there is the “Diary of Proficiency”, available in print format and use in the Online System (Education).
A new user always has doubts about the operation of the service and its responsibilities. For this reason, the first orientation is to read the Participant Manual, which explains how the program works, and the “Questions and Answers”, which clarify the participants’ most common doubts.
The main care is to run the control correctly so that the evaluation represents exclusively the performance of the routine and so that it is not contaminated with errors related to the quality control itself.
The main precautions include:
1. Read and follow the instructions for using the materials to avoid misunderstandings in the handling, since there are some procedures that differ from those adopted in routine patients;
2. Reconstitute freeze dried material with reagent water and with calibrated pipettor support, paying attention to the volume indicated on the label, since each material has a specific volume and errors in this step cause damage to the evaluation;
3. Use reconstituted material immediately to avoid variations resulting from improper storage;
4. Give special attention to dilutions and consequent calculations to avoid reporting erroneous results;
5. Check the need to convert results to the unit adopted by the program and perform the conversion correctly (see “Why do I have to convert my results to the unit adopted in the PT?”);
6. Identify the analytical system correctly (reagent, method, equipment, etc.) and report all requested information. According to the test, the absence of a data referring to the methodology may make its evaluation unfeasible;
7. Be aware of the response deadline and e-mail warnings (see “What happens when I stop responding to a round?”);
8. Contact Controllab immediately if you have any questions.
Although the modules are quarterly, participants with monthly payments split the amount. Thus, the invoice describes all the modules in which the participant is enrolled, and is charged the amount corresponding to the monthly apportionment. This is done so that the laboratory has a value for payment always equal, with greater ease of control.
Operation:
It’s simple! Check if the related form has the option to report this situation in the “+observation” field.
Form with the option “+observation”: select the alternative compatible with the laboratory scenario “Not performed: Not offered” or “Not performed: Outsource”. When these alternatives are checked, they are continuously applied in the form and the field for the test results reporting is disabled in the informed round and in the following rounds. To enable the results reporting for the marked test again, it is necessary to unmark the observation and apply.
Form without the option “+observation”: the laboratory must not respond to the test not performed or outsourced. You must leave the field blank.
As the modules of the PT (proficiency testing) are composed of test groups, it is foreseen that one test or another is not part of the participant’s routine.
If you outsource, do not send results that were not obtained by your laboratory, or, otherwise, it will appear in your evaluation and certificate as if it had been performed by your laboratory.
Each participant should only report results obtained in his routine, should not report tests performed by third parties or other technical unit of the laboratory. Just as you should not exchange information with other laboratories. This is because:
1. The objective of participating in the program is to have the opportunity to identify the good progress or not of each participant processes. When the results are obtained by another, there is no benefit.
2. When duplicating the result of a single laboratory, a trend is inserted in the result group that is not real, which may prejudice the statistical analysis and the evaluation of the other participants. The statistical model adopted presupposes only one result of each participant, which individually contributes to the determination of the target value used for the calculation of the evaluation range.
3. The certificate of proficiency and other documents proving participation in the program are issued to the enrollee in the program, which means that all the tests and results are executed by the enrollee.
The modules of the PT (proficiency testing) are quarterly to meet statistical requirements and ANVISA/REBLAS. The purpose of the proficiency testing is to identify errors, especially systemic ones. However, it is only possible to conclude whether an error is systemic (repetitive) or random, when at least two different items are tested.
With this, the first prerequisite is: never send a single sample per round. If we chose to submit monthly, we would have at least 24 items per year for each trial/module, and this would make the program much more costly.
On the other hand, there are no effective gains that justify such frequency and quantity of materials. Globally, 10 to 20 test items are adopted, with a common periodicity of 2 to 4 months.
But this rule presupposes that the laboratory adopt effective internal controls with more frequency in its routine (daily, between batches etc). Do not change internal control with external control. The first one is to verify its reproducibility and small deviations; already the proficiency test is mainly directed towards accuracy.
The rounds are sent according to the annual calendar, previously defined and distributed to the participants. The rounds are usually sent in the first half of the month, with some adaptations for the current year and holidays. Therefore, it is necessary the annual calendar available on the internet (website and online system).
Users of the Online System also receive an email notification of the round referral.
Not necessarily. Each module has its materials (test items), which can be:
• unique to more than one module (such as Biochemistry and Basic Hormones: all tests are dosed on the same item);
• unique for all tests of a module (as in Tumor Markers: an item for analyzing all markers);
• Individuals per test (as in Molecular Biology: one item for each type of test – HBV, HCV, and HIV);
• Different for test groups (such as in Cell Count: one material for global counting and one for differential);
• Items are labeled with an individual identification (e.g. BA01), and response forms (printed or online) repeat this identification for the tests to which the item corresponds.
This distribution is always the same and repeated every round, so that the participant becomes familiar with the system. Any change is communicated to the participants.
Results Report:
In order to make a statistical study and compare its results with those of the other participants of the PT (proficiency testing), it is necessary to adopt a single unit of measurement. Thus, we always seek to adopt the unit most used in the market.
To compare the results of different participants and to promote statistical calculations, it is necessary to set a default number of decimal places. Thus, a number of decimal places are chosen and/or relevant for the analysis of results.
The parameters calculated manually from multiple systems (using other parameters not directly related to the analysis at issue) are evaluated in an educational way in the program. These parameters do not present values obtained directly from the dosages, accumulate uncertainties due to the different variability contributions of the raw results and do not have the reliability required for an evaluation.
This “non-achievement – NR” counts towards your evaluation of the PT (Proficiency Testing) and must be recovered at the end of the period, with the special round. But attention! For tests that are not routinely performed, the process is different. See the answer to “How should I proceed with tests that I don´t perform/outsource”.
The rounds are periodic (quarterly, half yearly) and have a reduced number of test items (when compared to the volume of patients treated by the laboratory). Therefore, it is very important that the laboratory responds to all rounds and tests in its routine.
A non-participation is only paid when the laboratory has stopped activities, such as instrument breakage or lack of kit. In this case, it is necessary to inform Controllab. This allowance can only occur once a year, to maintain the annual representativeness of its results. See response to “If you can’t answer for lack of kits or broken equipment, what happens?”.
If, during the entire period for performing a round, the laboratory has its activities stopped due to operational problems (such as equipment downtime or lack of kit), check if the form related to the inoperative analytical system has the option to report this situation. in the “+ observation” field.
Form with option “+observation”: select the alternative “No inputs at the moment or EQU in maintenance/unavailable”. When this alternative is checked, the request is processed and the NPJ application will be automatically released along with the evaluation process.
Form without option “+observation”: the laboratory must contact Controllab to register the No Justified Participation (NPJ) request.
Thus, Controllab will be able to deduct this round from the laboratory’s score. However, this procedure cannot occur more than once a year, as the sampling of the Proficiency Testing would no longer be representative for the certificate.
Evaluation:
Evaluations are released an average of three weeks after the deadline for sending the results. New policies are being studied and implemented at all times, to increasingly reduce this evaluation period.
The evaluation process includes (1) receipt of the results, (2) statistical and technical analysis, (3) customer feedback analysis and (4) generation / access to evaluations.
In receipt of results, we are increasingly restricting the granting of increased response time to minimize the impact on the release of evaluations. In the analysis phase, the process has been optimized without giving up the analysis of qualified professionals and consulting the advisors in addition to the statistical treatment. These policies are to ensure a more thorough analysis and more reliable evaluations.
Customer feedback analysis occurs simultaneously with the statistical and technical analysis of the data and must be done completely for the release of the evaluation, since it may contain important information to define the evaluation. However, the volume of comments grows every day (ten to twenty times greater than when implementing the online system), with different information (see ‘When I comment on response forms do you respond?’) and often requires a lot of time in the evaluation process. Actions for the best use of the ‘feedback’ field and reduction of frequent questions have been adopted to minimize this impact.
The generation and access of the evaluation are already immediate since 2004. Once released, it is available to consultation in the Online System.
A delay in the evaluation is usually associated with some atypical performance in the data that requires a more complex analysis, such as material reanalysis (material quality control), contact with users / manufacturers or conducting research and consultation with third parties (other advisors).
Quality control materials are previously analyzed. They go through a quality control that aims to approve them regarding homogeneity and stability. Meanwhile, the evaluation of the participants’ results is the statistical model most used by the proficiency testing providers. This is because it has a consistent volume of results to evaluate accuracy and allows its comparison in a similar analytical system, since the control made by the provider does not cover all existing analytical systems.
However, it is important to note that a comparison is made between participants’ results and quality control to release an evaluation. Controllab always seeks consistency of these results so that there are no mistaken evaluations.
A result may not be evaluated because:
(a) didn’t form an evaluation group (less than 5 results);
(b) the evaluation group presented high variation;
(c) by determination of the advisory group, this decision is described in the document “Results Profile”.
In some specific situations, it is possible the evaluation in groups with less than 5 results, as statistical consistency of the data is verified.
There are a few possibilities:
1. Test with high variation: the results of the participants do not form a consistent statistic, presenting high variation. The option is to not evaluate based on this data.
2. Tests with predictable results: some analytes are only available (human matrix etc) in normal concentrations, and it is not possible to change them synthetically. In this case, they do not meet the requirement of proficiency test of “materials with unknown results” and no evaluation is done.
3. Insufficient number of evaluations: 8 to 16 items per year are sent according to the test, if more than 50% of the tests have not been evaluated (see the question “Why sometimes my results are not evaluated?”), the option is to not evaluate because the volume of evaluations is not representative for the year.
Certificate and related:
The certificate lists all tests with performance higher than or equal to that defined by ANVISA, in procedure GGLAS 02/43. We define these tests as “adequate” because they have reached the minimum acceptable level of performance.
The certificate also presents educational tests (see ‘Why are some tests classified as EDU?’). Inadequate tests, with less than acceptable performance, is not listed on the certificate. The “Management Report”, available for consultation in the Online System, maintains an updated list of inadequate tests.
It monitors systems analytical performance with each routine.
Internal Control
A reconstituição diária do CI (controle interno) evita a propagação de erros de reconstituição e possibilita uma estabilidade com menor dependência das condições de armazenagem. Por isso, reconstituir diariamente, sem dividir em alíquotas é a melhor forma de usá-lo.
Após abertura e reconstituição, o material de controle está sujeito a variações e contaminação, da mesma forma que os materiais de paciente. Assim, o usuário que optar pela aliquotagem, deve ter muito cuidado com a reconstituição e o armazenamento do material.
O uso de dois níveis de controle já vem sendo preconizado nas BPLC – Boas Práticas de Laboratórios Clínicos, na CLIA’88 (lei americana) e em processos de acreditação. Isto porque estudos apontam para uma maior eficiência do processo de controle com múltiplos níveis, que permitem maior rastreabilidade do processo e maior cobertura da faixa de medição.
O uso de apenas um nível de controle, basicamente obriga a rejeição, quando um resultado de controle excede 2DP,por exemplo. Isto leva a 5% de falsas rejeições para 1 leitura diária, 10% para 2 leituras diárias, e assim sucessivamente. E cada falsa rejeição gera retrabalho, custos, e não ajuda a conhecer melhor o processo nem a melhorá-lo.
O próprio Westgard disponibiliza um software que é capaz de fazer simulações estatísticas e definir a melhor combinação de níveis, freqüência e regras de controle para cada ensaio. Lembrando que, para alguns ensaios, a quantidade de níveis recomendada é superior a 2.
Assim, os fabricantes costumam disponibilizar seus controles com a quantidade de níveis mais apropriada para a amplitude da faixa de medição e de acordo com as características da metodologia empregada.
Monitors laboratory requirements and improves operational results.
First Steps to Implementation:
The laboratory indicators are essential elements in the practice of measurement of management and process performance, allowing a comparative evaluation, at a national and international level, of the results of the clinical laboratories in front of their peers and market leaders.
This process of benchmarking of practices and metrics, also known as Benchmarking, allows laboratories to be more effective in decision-making against their strategies and encourages the continuous improvement of their processes, contributing to patient safety and to increasing productivity and sustainability of the laboratory medicine sector.
A new user must first define which data the laboratory is already able to monitor and mark this data in the “Manage Indicators” menu. With this selection, the system will automatically display which indicators were enabled to obtain the benchmarking.
In order for the indicators to contribute to improving the quality of health care, it is necessary that laboratory staff involved in the monitored processes have access to data specifications and how they should be collected, especially in the critical aspects of the laboratory process.
If you have any questions, you should contact Controllab.
Operation:
The discussion group is a group of volunteers from a number of different laboratories and is responsible for defining and detailing the indicators, analyzing overall results, clarifying questions and proposing improvements to the program.
Yes. The data is analyzed as a whole without verification from the laboratory that sent us and, for greater security, the laboratories are given an access number so that their name is not exposed in any material. All data available for other laboratories are general data.
Comparison Between Networks (support or brands):
In “My Networks” the user can create their own comparison network for a closed benchmarking. In this tool, he may invite support networks or brand labs to compare their performance with each other.
For the comparison to be generated, network members must respond within the reporting period of the indicators. The member who does not respond to the indicator will not have access to the comparison.
To create a comparison network, the Laboratory Indicator Benchmarking Program participant must access the option “My Networks†(Online System> Indicators Benchmarking> My Networks> Create New Network).
To invite other labs you will need the name of the institution, the name of the person in charge and the contact email. Once the information is completed and the network has been created, all invited members will receive an activation code to confirm acceptance in the registered emails.
Invited members will receive an activation code in their registered emails. This code must be entered in the Online System > Control Center > My Networks > Enter Code. Once the code is entered and the action confirmed, this user will be a member of the network.
The invited laboratory not participating in the Indicators Benchmarking Program should contact our Call Center (+55 21 3891-9900 – atendimento@controllab.com).
Any laboratory participating in the Controllab Laboratory Benchmarking Program. However, the invitation may only be sent and accepted by the user registered as an “administrator” profile in the Online System.
The network is active only when all invited members accept the invitation. If any invitations are still pending, the network status will also be pending. In this case, the institution responsible for the created network may resend the invitation or delete this member. Once the network is activated, the comparison will be generated within the same period as the Program benchmarking release (on the first day of each month).
There are 3 identification options:
- Total: All institutions can be identified in the reports.
- Partial: Only the institution that created the network can identify the others in the reports.
- Locked: No institutions will be identified in the report (minimum 3 institutions required).
During the creation of the network, the user must choose one of these options and, in the invitation to the laboratories, the identification option configured by the responsible institution will be displayed.
Only the institution responsible for the network (institution you created from the comparison network) can manage. However, any member of the group (with the profile “administrator” in the Online System) can leave your comparison network at any time.
From the exclusion date, the indicators will no longer be compared between network members. Once the group is deleted, it cannot be reactivated.
Optimize your routine resources.
Strain Control
The certificates are made available online. They are digitally signed, ensuring the authenticity of the document.
Users registered for monthly deliveries receive a login and password to access the Online System. In this system, certificates in PDF file are available in the Internal Control menu, while the material lasts. Users who purchase the materials separately, receive the link with the certificate file via email.
Some PDF settings may generate signature invalidity messages or similar. In these cases, we ask you to adjust the PDF settings illustrated in this video.
Insert more accuracy anda traceability to the analytical process.
Certified Reference Materials
The certificates are made available online. They are digitally signed, ensuring the authenticity of the document.
Users registered for monthly deliveries receive a login and password to access the Online System. In this system, certificates in PDF file are available in the solution menu, while the material lasts. Users who purchase the materials separately, receive the link with the certificate file via email.
Some PDF settings may generate signature invalidity messages or similar. In these cases, we ask you to adjust the PDF settings illustrated in this video.
Prevents measurement errors in analysis.
Calibração de Instrumentos
Não. A calibração é um processo de medição que informa o volume real dispensado ou contido por uma vidraria, a medida real de um termômetro com relação ao indicado na sua graduação etc. Em resumo, é um “raio x” do funcionamento atual do instrumento.
Como resultado da calibração, é possível decidir se o instrumento está adequado ao uso. Alguns instrumentos, ao apresentarem desvio frente ao esperado (erro), podem ser ajustados, desde que o custo do conserto justifique e possuam mecanismos para o ajuste.
Cabe ao usuário definir qual o erro máximo permitido, para que a empresa responsável pela manutenção faça o ajuste. Realizado o ajuste, é necessário efetuar uma nova calibração para verificar sua eficácia.
Não existe uma “validade” previamente definida para um certificado de calibração. O usuário é quem deve definir o intervalo entre calibrações a ser adotado para cada instrumento, considerando a sua freqüência de uso e os resultados obtidos nas últimas calibrações.
Inicialmente, quando não há um histórico do instrumento, é recomendado o uso de prazos curtos. Isto evita que um instrumento fora das condições adequadas impacte um grande volume de resultados por um longo tempo, permitindo ação corretiva mais rápida e menor impacto na rotina.
O certificado de calibração apresenta a média (obtida pelo instrumento após sucessivas medidas) e a sua incerteza (variação das medidas). O usuário deve verificar se o intervalo de resultados (média ± incerteza) apresentado no certificado está contido no erro máximo permitido pelo laboratório (normalmente chamado de tolerância).
Em alguns instrumentos, isto é possível; porém, não é recomendado.
As micropipetas/pipetadores costumam possuir uma trava interna, que pode ser violada pelo usuário, ao forçar o ajuste para volumes acima ou abaixo da faixa de uso recomendada.
Contudo, essa prática pode iniciar um processo de perda da regulagem e da calibração, além de soltar ou quebrar a rosca, fazendo com que o instrumento passe a dispensar volumes errados. É necessário também muito cuidado com instrumentos cujos valores indicados no volumétrico são multiplicados por 10 para a leitura do valor dispensado.