A common question among science consumers today is the fact that researchers are publishing a lot, and a lot fast. To have a mathematical dimension, every nine years, the production scientific doubles its size [1]. The growing number of publications in all the areas, especially in the health area, combined with the ease of access to scientific documents, allows current professionals to guide their practice in evidence, in the hope of improving the quality of their care [1,2]. To make decisions, have scientific knowledge as a basis for clinical reasoning provides the health professional with an aura of safety and efficiency [2].However, the consumption of a scientific article can hide many tattoos that appear before reading the document. The so-called cognitive biases are inherent to human behavior and directly influence the understanding of literature, distorting conceptual reality. These work as a varnish to legitimize and strengthen the personal beliefs of the reader, who comes to believe that he is making the best decision based on scientific evidence [3,4]. Several aspects can hinder the interpretation and correct application of evidence-based practice concepts. The purpose of this document is to call attention to some of the cognitive biases, the importance of critical reading, and to present a flowchart that will assist the reader in making decisions about his practice.The term Cognitive Bias was coined by researchers Amos Tversky and Daniel Kahneman for the first time in 1972. Since then, several researchers have described different types of biases that affect decision-making in a wide range of areas of knowledge, including health [3]. We can exemplify the tendency in human behavior using as an archetype a person who, in possession of a resource (equipment, technique, method, drug), observed a positive response in two or three peoples submitted to his intervention. The positive feeling about the intervention (affinity bias) will make him believe that the change in the outcome has as major cause the applied intervention. When challenged about the veracity of his intervention, this professional tends to look for "evidence" that is generally supported by people who have already undergone the intervention and have benefited, and the search for scientific literature that often happens selectively (literature selection bias) to validate your belief (confirmation bias) [4].This illustration brings us fundamental points for understanding the real meaning of scientific thought. It is necessary to assume that, as professionals, we have preferences and that they can cloud our judgment, this recognition and attempt at impartiality being the beginning for the most appropriate acceptance of scientific evidence. In the background, it is necessary to understand that these same biases are also present in the authors of the research. Therefore, on purpose or not, scientific writing carries, to a greater or lesser degree, the author's preferences. It is up to the reader to identify which ideas are or are not supported by overwhelming data.In the above example, the absence of a control group prevents us from measuring the benefit objectively. In this situation, the assessment of a subjective outcome, such as quality of life, can lead to a false-positive result due to the placebo effect. Other measures may occur depending on the time. For these reasons, clinical practice observations are subject to considerable risks of bias and interpretation, as it is an uncontrolled environment.Still, on the biases of human understanding, a search for positive results for a specific intervention reduces the number of treatment possibilities and harm the final acceptor of the health system, our patient. In the field of scientific literature, it is also possible to identify data that justify dichotomous thoughts. At this point, it is necessary to emphasize the importance of understanding and differentiating each study design, the type of inference allowed, and the analysis of biases present in the research, which will give us to trust or not the result, especially when there is a positive response to the intervention [5]. The flaws in the comparisons between the evidence can lead to the choice of misconduct (not supported by substantiated results) and influence by an affinity bias.Regarding the proof of results, a special warning is necessary: can be shown the significance of an outcome from a statistical and clinical point of view. In the first case, one must pay attention to the statistical test applied, the nature of the comparisons, and the presence of comparable groups at the beginning of an intervention study, when possible. The application of the statistical test, the assembly of the sample calculation, and the change in the outcome predictor variable can be used to achieve a statistically significant result, which unfortunately is still the most accepted result in scientific journals, and act against the scientific thinking of quality [6].Even in the presence of statistical findings, we need to reflect on the clinical importance of this finding [7]. Let's imagine conducting a study using an exercise protocol aimed at responding to systolic blood pressure (SBP) in hypertensive men. Both groups of hypertensive patients had an SBP of 170mmHg at the beginning of the study, and we found that the intervention group and the control group after three months had SBP equal to 168mmHg and 169mmHg, respectively. Checking these data, followed by a statistically significant statement (p