Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 28.328
Filtrar
Mais filtros

Coleções SMS-SP
Intervalo de ano de publicação
1.
Cell ; 174(6): 1424-1435.e15, 2018 09 06.
Artigo em Inglês | MEDLINE | ID: mdl-30078708

RESUMO

FOXP2, initially identified for its role in human speech, contains two nonsynonymous substitutions derived in the human lineage. Evidence for a recent selective sweep in Homo sapiens, however, is at odds with the presence of these substitutions in archaic hominins. Here, we comprehensively reanalyze FOXP2 in hundreds of globally distributed genomes to test for recent selection. We do not find evidence of recent positive or balancing selection at FOXP2. Instead, the original signal appears to have been due to sample composition. Our tests do identify an intronic region that is enriched for highly conserved sites that are polymorphic among humans, compatible with a loss of function in humans. This region is lowly expressed in relevant tissue types that were tested via RNA-seq in human prefrontal cortex and RT-PCR in immortalized human brain cells. Our results represent a substantial revision to the adaptive history of FOXP2, a gene regarded as vital to human evolution.


Assuntos
Fatores de Transcrição Forkhead/genética , Encéfalo/citologia , Encéfalo/metabolismo , Linhagem Celular , Bases de Dados Genéticas , Éxons , Feminino , Genoma Humano , Haplótipos , Humanos , Íntrons , Masculino , Cadeias de Markov , Polimorfismo de Nucleotídeo Único , Córtex Pré-Frontal/metabolismo
2.
Annu Rev Neurosci ; 47(1): 277-301, 2024 Aug.
Artigo em Inglês | MEDLINE | ID: mdl-38669478

RESUMO

It has long been argued that only humans could produce and understand language. But now, for the first time, artificial language models (LMs) achieve this feat. Here we survey the new purchase LMs are providing on the question of how language is implemented in the brain. We discuss why, a priori, LMs might be expected to share similarities with the human language system. We then summarize evidence that LMs represent linguistic information similarly enough to humans to enable relatively accurate brain encoding and decoding during language processing. Finally, we examine which LM properties-their architecture, task performance, or training-are critical for capturing human neural responses to language and review studies using LMs as in silico model organisms for testing hypotheses about language. These ongoing investigations bring us closer to understanding the representations and processes that underlie our ability to comprehend sentences and express thoughts in language.


Assuntos
Encéfalo , Idioma , Humanos , Encéfalo/fisiologia , Animais , Inteligência Artificial , Modelos Neurológicos
3.
Mol Cell ; 84(7): 1257-1270.e6, 2024 Apr 04.
Artigo em Inglês | MEDLINE | ID: mdl-38377993

RESUMO

Current base editors (BEs) use DNA deaminases, including cytidine deaminase in cytidine BE (CBE) or adenine deaminase in adenine BE (ABE), to facilitate transition nucleotide substitutions. Combining CBE or ABE with glycosylase enzymes can induce limited transversion mutations. Nonetheless, a critical demand remains for BEs capable of generating alternative mutation types, such as T>G corrections. In this study, we leveraged pre-trained protein language models to optimize a uracil-N-glycosylase (UNG) variant with altered specificity for thymines (eTDG). Notably, after two rounds of testing fewer than 50 top-ranking variants, more than 50% exhibited over 1.5-fold enhancement in enzymatic activities. When eTDG was fused with nCas9, it induced programmable T-to-S (G/C) substitutions and corrected db/db diabetic mutation in mice (up to 55%). Our findings not only establish orthogonal strategies for developing novel BEs but also demonstrate the capacities of protein language models for optimizing enzymes without extensive task-specific training data.


Assuntos
Ácidos Alcanossulfônicos , Edição de Genes , Uracila-DNA Glicosidase , Animais , Camundongos , Mutação , Uracila-DNA Glicosidase/genética , Uracila-DNA Glicosidase/metabolismo
4.
Cell ; 164(6): 1269-1276, 2016 Mar 10.
Artigo em Inglês | MEDLINE | ID: mdl-26967292

RESUMO

The use of vocalizations to communicate information and elaborate social bonds is an adaptation seen in many vertebrate species. Human speech is an extreme version of this pervasive form of communication. Unlike the vocalizations exhibited by the majority of land vertebrates, speech is a learned behavior requiring early sensory exposure and auditory feedback for its development and maintenance. Studies in humans and a small number of other species have provided insights into the neural and genetic basis for learned vocal communication and are helping to delineate the roles of brain circuits across the cortex, basal ganglia, and cerebellum in generating vocal behaviors. This Review provides an outline of the current knowledge about these circuits and the genes implicated in vocal communication, as well as a perspective on future research directions in this field.


Assuntos
Fala , Vocalização Animal , Animais , Encéfalo/fisiologia , Fatores de Transcrição Forkhead/genética , Fatores de Transcrição Forkhead/metabolismo , Humanos , Aprendizagem , Doenças do Sistema Nervoso/genética , Vias Neurais
5.
Annu Rev Neurosci ; 45: 295-316, 2022 07 08.
Artigo em Inglês | MEDLINE | ID: mdl-35316612

RESUMO

Vocal communication is a critical feature of social interaction across species; however, the relation between such behavior in humans and nonhumans remains unclear. To enable comparative investigation of this topic, we review the literature pertinent to interactive language use and identify the superset of cognitive operations involved in generating communicative action. We posit these functions comprise three intersecting multistep pathways: (a) the Content Pathway, which selects the movements constituting a response; (b) the Timing Pathway, which temporally structures responses; and (c) the Affect Pathway, which modulates response parameters according to internal state. These processing streams form the basis of the Convergent Pathways for Interaction framework, which provides a conceptual model for investigating the cognitive and neural computations underlying vocal communication across species.


Assuntos
Idioma , Vocalização Animal , Animais , Humanos , Vocalização Animal/fisiologia
6.
Trends Biochem Sci ; 48(12): 1014-1018, 2023 12.
Artigo em Inglês | MEDLINE | ID: mdl-37833131

RESUMO

Generative artificial intelligence (AI) is a burgeoning field with widespread applications, including in science. Here, we explore two paradigms that provide insight into the capabilities and limitations of Chat Generative Pre-trained Transformer (ChatGPT): its ability to (i) define a core biological concept (the Central Dogma of molecular biology); and (ii) interpret the genetic code.


Assuntos
Inteligência Artificial , Código Genético , Biologia Molecular
7.
Physiol Rev ; 100(3): 1019-1063, 2020 07 01.
Artigo em Inglês | MEDLINE | ID: mdl-32233912

RESUMO

Comparative studies on brain asymmetry date back to the 19th century but then largely disappeared due to the assumption that lateralization is uniquely human. Since the reemergence of this field in the 1970s, we learned that left-right differences of brain and behavior exist throughout the animal kingdom and pay off in terms of sensory, cognitive, and motor efficiency. Ontogenetically, lateralization starts in many species with asymmetrical expression patterns of genes within the Nodal cascade that set up the scene for later complex interactions of genetic, environmental, and epigenetic factors. These take effect during different time points of ontogeny and create asymmetries of neural networks in diverse species. As a result, depending on task demands, left- or right-hemispheric loops of feedforward or feedback projections are then activated and can temporarily dominate a neural process. In addition, asymmetries of commissural transfer can shape lateralized processes in each hemisphere. It is still unclear if interhemispheric interactions depend on an inhibition/excitation dichotomy or instead adjust the contralateral temporal neural structure to delay the other hemisphere or synchronize with it during joint action. As outlined in our review, novel animal models and approaches could be established in the last decades, and they already produced a substantial increase of knowledge. Since there is practically no realm of human perception, cognition, emotion, or action that is not affected by our lateralized neural organization, insights from these comparative studies are crucial to understand the functions and pathologies of our asymmetric brain.


Assuntos
Evolução Biológica , Encéfalo/fisiologia , Lateralidade Funcional/genética , Lateralidade Funcional/fisiologia , Animais , Encéfalo/anatomia & histologia , História do Século XIX , História do Século XX , História do Século XXI , Humanos , Pesquisa/história
8.
Am J Hum Genet ; 2024 Jul 31.
Artigo em Inglês | MEDLINE | ID: mdl-39146935

RESUMO

Large language models (LLMs) are generating interest in medical settings. For example, LLMs can respond coherently to medical queries by providing plausible differential diagnoses based on clinical notes. However, there are many questions to explore, such as evaluating differences between open- and closed-source LLMs as well as LLM performance on queries from both medical and non-medical users. In this study, we assessed multiple LLMs, including Llama-2-chat, Vicuna, Medllama2, Bard/Gemini, Claude, ChatGPT3.5, and ChatGPT-4, as well as non-LLM approaches (Google search and Phenomizer) regarding their ability to identify genetic conditions from textbook-like clinician questions and their corresponding layperson translations related to 63 genetic conditions. For open-source LLMs, larger models were more accurate than smaller LLMs: 7b, 13b, and larger than 33b parameter models obtained accuracy ranges from 21%-49%, 41%-51%, and 54%-68%, respectively. Closed-source LLMs outperformed open-source LLMs, with ChatGPT-4 performing best (89%-90%). Three of 11 LLMs and Google search had significant performance gaps between clinician and layperson prompts. We also evaluated how in-context prompting and keyword removal affected open-source LLM performance. Models were provided with 2 types of in-context prompts: list-type prompts, which improved LLM performance, and definition-type prompts, which did not. We further analyzed removal of rare terms from descriptions, which decreased accuracy for 5 of 7 evaluated LLMs. Finally, we observed much lower performance with real individuals' descriptions; LLMs answered these questions with a maximum 21% accuracy.

9.
Proc Natl Acad Sci U S A ; 121(10): e2307876121, 2024 Mar 05.
Artigo em Inglês | MEDLINE | ID: mdl-38422017

RESUMO

During real-time language comprehension, our minds rapidly decode complex meanings from sequences of words. The difficulty of doing so is known to be related to words' contextual predictability, but what cognitive processes do these predictability effects reflect? In one view, predictability effects reflect facilitation due to anticipatory processing of words that are predictable from context. This view predicts a linear effect of predictability on processing demand. In another view, predictability effects reflect the costs of probabilistic inference over sentence interpretations. This view predicts either a logarithmic or a superlogarithmic effect of predictability on processing demand, depending on whether it assumes pressures toward a uniform distribution of information over time. The empirical record is currently mixed. Here, we revisit this question at scale: We analyze six reading datasets, estimate next-word probabilities with diverse statistical language models, and model reading times using recent advances in nonlinear regression. Results support a logarithmic effect of word predictability on processing difficulty, which favors probabilistic inference as a key component of human language processing.


Assuntos
Compreensão , Idioma , Humanos , Modelos Estatísticos
10.
Proc Natl Acad Sci U S A ; 121(11): e2310766121, 2024 Mar 12.
Artigo em Inglês | MEDLINE | ID: mdl-38442171

RESUMO

The neural correlates of sentence production are typically studied using task paradigms that differ considerably from the experience of speaking outside of an experimental setting. In this fMRI study, we aimed to gain a better understanding of syntactic processing in spontaneous production versus naturalistic comprehension in three regions of interest (BA44, BA45, and left posterior middle temporal gyrus). A group of participants (n = 16) was asked to speak about the events of an episode of a TV series in the scanner. Another group of participants (n = 36) listened to the spoken recall of a participant from the first group. To model syntactic processing, we extracted word-by-word metrics of phrase-structure building with a top-down and a bottom-up parser that make different hypotheses about the timing of structure building. While the top-down parser anticipates syntactic structure, sometimes before it is obvious to the listener, the bottom-up parser builds syntactic structure in an integratory way after all of the evidence has been presented. In comprehension, neural activity was found to be better modeled by the bottom-up parser, while in production, it was better modeled by the top-down parser. We additionally modeled structure building in production with two strategies that were developed here to make different predictions about the incrementality of structure building during speaking. We found evidence for highly incremental and anticipatory structure building in production, which was confirmed by a converging analysis of the pausing patterns in speech. Overall, this study shows the feasibility of studying the neural dynamics of spontaneous language production.


Assuntos
Benchmarking , Rememoração Mental , Humanos , Idioma , Software , Fala
11.
Proc Natl Acad Sci U S A ; 121(30): e2315438121, 2024 Jul 23.
Artigo em Inglês | MEDLINE | ID: mdl-39028693

RESUMO

There is evidence from both behavior and brain activity that the way information is structured, through the use of focus, can up-regulate processing of focused constituents, likely to give prominence to the relevant aspects of the input. This is hypothesized to be universal, regardless of the different ways in which languages encode focus. In order to test this universalist hypothesis, we need to go beyond the more familiar linguistic strategies for marking focus, such as by means of intonation or specific syntactic structures (e.g., it-clefts). Therefore, in this study, we examine Makhuwa-Enahara, a Bantu language spoken in northern Mozambique, which uniquely marks focus through verbal conjugation. The participants were presented with sentences that consisted of either a semantically anomalous constituent or a semantically nonanomalous constituent. Moreover, focus on this particular constituent could be either present or absent. We observed a consistent pattern: Focused information generated a more negative N400 response than the same information in nonfocus position. This demonstrates that regardless of how focus is marked, its consequence seems to result in an upregulation of processing of information that is in focus.


Assuntos
Idioma , Humanos , Feminino , Masculino , Adulto , Moçambique , Eletroencefalografia , Semântica , Encéfalo/fisiologia , Adulto Jovem , Linguística , Potenciais Evocados/fisiologia
12.
Proc Natl Acad Sci U S A ; 121(25): e2320066121, 2024 Jun 18.
Artigo em Inglês | MEDLINE | ID: mdl-38861605

RESUMO

How are the merits of innovative ideas communicated in science? Here, we conduct semantic analyses of grant application success with a focus on scientific promotional language, which may help to convey an innovative idea's originality and significance. Our analysis attempts to surmount the limitations of prior grant studies by examining the full text of tens of thousands of both funded and unfunded grants from three leading public and private funding agencies: the NIH, the NSF, and the Novo Nordisk Foundation, one of the world's largest private science funding foundations. We find a robust association between promotional language and the support and adoption of innovative ideas by funders and other scientists. First, a grant proposal's percentage of promotional language is associated with up to a doubling of the grant's probability of being funded. Second, a grant's promotional language reflects its intrinsic innovativeness. Third, the percentage of promotional language is predictive of the expected citation and productivity impact of publications that are supported by funded grants. Finally, a computer-assisted experiment that manipulates the promotional language in our data demonstrates how promotional language can communicate the merit of ideas through cognitive activation. With the incidence of promotional language in science steeply rising, and the pivotal role of grants in converting promising and aspirational ideas into solutions, our analysis provides empirical evidence that promotional language is associated with effectively communicating the merits of innovative scientific ideas.


Assuntos
Idioma , Humanos , Ciência , Organização do Financiamento , Estados Unidos , Apoio à Pesquisa como Assunto , Criatividade
13.
Proc Natl Acad Sci U S A ; 121(24): e2317967121, 2024 Jun 11.
Artigo em Inglês | MEDLINE | ID: mdl-38833474

RESUMO

Large language models (LLMs) are currently at the forefront of intertwining AI systems with human communication and everyday life. Thus, aligning them with human values is of great importance. However, given the steady increase in reasoning abilities, future LLMs are under suspicion of becoming able to deceive human operators and utilizing this ability to bypass monitoring efforts. As a prerequisite to this, LLMs need to possess a conceptual understanding of deception strategies. This study reveals that such strategies emerged in state-of-the-art LLMs, but were nonexistent in earlier LLMs. We conduct a series of experiments showing that state-of-the-art LLMs are able to understand and induce false beliefs in other agents, that their performance in complex deception scenarios can be amplified utilizing chain-of-thought reasoning, and that eliciting Machiavellianism in LLMs can trigger misaligned deceptive behavior. GPT-4, for instance, exhibits deceptive behavior in simple test scenarios 99.16% of the time (P < 0.001). In complex second-order deception test scenarios where the aim is to mislead someone who expects to be deceived, GPT-4 resorts to deceptive behavior 71.46% of the time (P < 0.001) when augmented with chain-of-thought reasoning. In sum, revealing hitherto unknown machine behavior in LLMs, our study contributes to the nascent field of machine psychology.


Assuntos
Enganação , Idioma , Humanos , Inteligência Artificial
14.
Proc Natl Acad Sci U S A ; 121(32): e2402068121, 2024 Aug 06.
Artigo em Inglês | MEDLINE | ID: mdl-39088395

RESUMO

Linguistic communication is an intrinsically social activity that enables us to share thoughts across minds. Many complex social uses of language can be captured by domain-general representations of other minds (i.e., mentalistic representations) that externally modulate linguistic meaning through Gricean reasoning. However, here we show that representations of others' attention are embedded within language itself. Across ten languages, we show that demonstratives-basic grammatical words (e.g., "this"/"that") which are evolutionarily ancient, learned early in life, and documented in all known languages-are intrinsic attention tools. Beyond their spatial meanings, demonstratives encode both joint attention and the direction in which the listener must turn to establish it. Crucially, the frequency of the spatial and attentional uses of demonstratives varies across languages, suggesting that both spatial and mentalistic representations are part of their conventional meaning. Using computational modeling, we show that mentalistic representations of others' attention are internally encoded in demonstratives, with their effect further boosted by Gricean reasoning. Yet, speakers are largely unaware of this, incorrectly reporting that they primarily capture spatial representations. Our findings show that representations of other people's cognitive states (namely, their attention) are embedded in language and suggest that the most basic building blocks of the linguistic system crucially rely on social cognition.


Assuntos
Atenção , Idioma , Humanos , Atenção/fisiologia , Cognição/fisiologia , Linguística , Comunicação , Feminino , Masculino
15.
Proc Natl Acad Sci U S A ; 121(18): e2312323121, 2024 Apr 30.
Artigo em Inglês | MEDLINE | ID: mdl-38621117

RESUMO

Zebra finches, a species of songbirds, learn to sing by creating an auditory template through the memorization of model songs (sensory learning phase) and subsequently translating these perceptual memories into motor skills (sensorimotor learning phase). It has been traditionally believed that babbling in juvenile birds initiates the sensorimotor phase while the sensory phase of song learning precedes the onset of babbling. However, our findings challenge this notion by demonstrating that testosterone-induced premature babbling actually triggers the onset of the sensory learning phase instead. We reveal that juvenile birds must engage in babbling and self-listening to acquire the tutor song as the template. Notably, the sensory learning of the template in songbirds requires motor vocal activity, reflecting the observation that prelinguistic babbling in humans plays a crucial role in auditory learning for language acquisition.


Assuntos
Tentilhões , Animais , Humanos , Vocalização Animal , Aprendizagem , Desenvolvimento da Linguagem
16.
Proc Natl Acad Sci U S A ; 121(2): e2306286121, 2024 Jan 09.
Artigo em Inglês | MEDLINE | ID: mdl-38175869

RESUMO

Adult second language (L2) learning is a challenging enterprise inducing neuroplastic changes in the human brain. However, it remains unclear how the structural language connectome and its subnetworks change during adult L2 learning. The current study investigated longitudinal changes in white matter (WM) language networks in each hemisphere, as well as their interconnection, in a large group of Arabic-speaking adults who learned German intensively for 6 mo. We found a significant increase in WM-connectivity within bilateral temporal-parietal semantic and phonological subnetworks and right temporal-frontal pathways mainly in the second half of the learning period. At the same time, WM-connectivity between the two hemispheres decreased significantly. Crucially, these changes in WM-connectivity are correlated with L2 performance. The observed changes in subnetworks of the two hemispheres suggest a network reconfiguration due to lexical learning. The reduced interhemispheric connectivity may indicate a key role of the corpus callosum in L2 learning by reducing the inhibition of the language-dominant left hemisphere. Our study highlights the dynamic changes within and across hemispheres in adult language-related networks driven by L2 learning.


Assuntos
Substância Branca , Adulto , Humanos , Idioma , Encéfalo/fisiologia , Aprendizagem/fisiologia , Semântica , Imageamento por Ressonância Magnética
17.
Proc Natl Acad Sci U S A ; 121(23): e2311425121, 2024 Jun 04.
Artigo em Inglês | MEDLINE | ID: mdl-38814865

RESUMO

Theories of language development-informed largely by studies of Western, middleclass infants-have highlighted the language that caregivers direct to children as a key driver of language learning. However, some have argued that language development unfolds similarly across environmental contexts, including those in which childdirected language is scarce. This raises the possibility that children are able to learn from other sources of language in their environments, particularly the language directed to others in their environment. We explore this hypothesis with infants in an indigenous Tseltal-speaking community in Southern Mexico who are rarely spoken to, yet have the opportunity to overhear a great deal of other-directed language by virtue of being carried on their mothers' backs. Adapting a previously established gaze-tracking method for detecting early word knowledge to our field setting, we find that Tseltal infants exhibit implicit knowledge of common nouns (Exp. 1), analogous to their US peers who are frequently spoken to. Moreover, they exhibit comprehension of Tseltal honorific terms that are exclusively used to greet adults in the community (Exp. 2), representing language that could only have been learned through overhearing. In so doing, Tseltal infants demonstrate an ability to discriminate words with similar meanings and perceptually similar referents at an earlier age than has been shown among Western children. Together, these results suggest that for some infants, learning from overhearing may be an important path toward developing language.


Assuntos
Compreensão , Desenvolvimento da Linguagem , Humanos , Lactente , Feminino , Masculino , Compreensão/fisiologia , México , Idioma , Vocabulário
18.
Proc Natl Acad Sci U S A ; 121(24): e2318124121, 2024 Jun 11.
Artigo em Inglês | MEDLINE | ID: mdl-38830100

RESUMO

There is much excitement about the opportunity to harness the power of large language models (LLMs) when building problem-solving assistants. However, the standard methodology of evaluating LLMs relies on static pairs of inputs and outputs; this is insufficient for making an informed decision about which LLMs are best to use in an interactive setting, and how that varies by setting. Static assessment therefore limits how we understand language model capabilities. We introduce CheckMate, an adaptable prototype platform for humans to interact with and evaluate LLMs. We conduct a study with CheckMate to evaluate three language models (InstructGPT, ChatGPT, and GPT-4) as assistants in proving undergraduate-level mathematics, with a mixed cohort of participants from undergraduate students to professors of mathematics. We release the resulting interaction and rating dataset, MathConverse. By analyzing MathConverse, we derive a taxonomy of human query behaviors and uncover that despite a generally positive correlation, there are notable instances of divergence between correctness and perceived helpfulness in LLM generations, among other findings. Further, we garner a more granular understanding of GPT-4 mathematical problem-solving through a series of case studies, contributed by experienced mathematicians. We conclude with actionable takeaways for ML practitioners and mathematicians: models that communicate uncertainty, respond well to user corrections, and can provide a concise rationale for their recommendations, may constitute better assistants. Humans should inspect LLM output carefully given their current shortcomings and potential for surprising fallibility.


Assuntos
Idioma , Matemática , Resolução de Problemas , Humanos , Resolução de Problemas/fisiologia , Estudantes/psicologia
19.
Proc Natl Acad Sci U S A ; 121(34): e2401687121, 2024 Aug 20.
Artigo em Inglês | MEDLINE | ID: mdl-39133845

RESUMO

The language network of the human brain has core components in the inferior frontal cortex and superior/middle temporal cortex, with left-hemisphere dominance in most people. Functional specialization and interconnectivity of these neocortical regions is likely to be reflected in their molecular and cellular profiles. Excitatory connections between cortical regions arise and innervate according to layer-specific patterns. Here, we generated a gene expression dataset from human postmortem cortical tissue samples from core language network regions, using spatial transcriptomics to discriminate gene expression across cortical layers. Integration of these data with existing single-cell expression data identified 56 genes that showed differences in laminar expression profiles between the frontal and temporal language cortex together with upregulation in layer II/III and/or layer V/VI excitatory neurons. Based on data from large-scale genome-wide screening in the population, DNA variants within these 56 genes showed set-level associations with interindividual variation in structural connectivity between the left-hemisphere frontal and temporal language cortex, and with the brain-related disorders dyslexia and schizophrenia which often involve affected language. These findings identify region-specific patterns of laminar gene expression as a feature of the brain's language network.


Assuntos
Idioma , Neocórtex , Humanos , Neocórtex/metabolismo , Lobo Temporal/metabolismo , Masculino , Feminino , Esquizofrenia/genética , Esquizofrenia/metabolismo , Neurônios/metabolismo , Lobo Frontal/metabolismo , Transcriptoma , Adulto
20.
Proc Natl Acad Sci U S A ; 121(34): e2308950121, 2024 Aug 20.
Artigo em Inglês | MEDLINE | ID: mdl-39133853

RESUMO

The social and behavioral sciences have been increasingly using automated text analysis to measure psychological constructs in text. We explore whether GPT, the large-language model (LLM) underlying the AI chatbot ChatGPT, can be used as a tool for automated psychological text analysis in several languages. Across 15 datasets (n = 47,925 manually annotated tweets and news headlines), we tested whether different versions of GPT (3.5 Turbo, 4, and 4 Turbo) can accurately detect psychological constructs (sentiment, discrete emotions, offensiveness, and moral foundations) across 12 languages. We found that GPT (r = 0.59 to 0.77) performed much better than English-language dictionary analysis (r = 0.20 to 0.30) at detecting psychological constructs as judged by manual annotators. GPT performed nearly as well as, and sometimes better than, several top-performing fine-tuned machine learning models. Moreover, GPT's performance improved across successive versions of the model, particularly for lesser-spoken languages, and became less expensive. Overall, GPT may be superior to many existing methods of automated text analysis, since it achieves relatively high accuracy across many languages, requires no training data, and is easy to use with simple prompts (e.g., "is this text negative?") and little coding experience. We provide sample code and a video tutorial for analyzing text with the GPT application programming interface. We argue that GPT and other LLMs help democratize automated text analysis by making advanced natural language processing capabilities more accessible, and may help facilitate more cross-linguistic research with understudied languages.


Assuntos
Multilinguismo , Humanos , Idioma , Aprendizado de Máquina , Processamento de Linguagem Natural , Emoções , Mídias Sociais
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA