Views about ChatGPT: Are human decision making and human learning necessary?

Fiche du document

Date

6 juin 2024

Type de document
Périmètre
Identifiant
  • 2406.03823
Collection

arXiv

Organisation

Cornell University




Citer ce document

Eiji Yamamura et al., « Views about ChatGPT: Are human decision making and human learning necessary? », arXiv - économie


Partage / Export

Résumé 0

Using individual-level survey data from 2024, this study investigated how respondent characteristics are associated with a subjective view of generative artificial intelligence (GAI). We asked 14 questions concerning respondents view about GAI, such as general view, faulty GAI, autonomous GEI, GAI replacing humans, and importance of human learning. Regression analysis based on the ordered logit model revealed that: (1) In some cases, the results of smartphone and computer usage times differed. Smartphone usage time was negatively correlated with the importance of human learning, whereas computer usage was not negatively correlated. (2) Managers and ordinary businesspeople have positive views of GAI. However, managers do not show a positive view about GAI being responsible for human decision making. (3) Teachers generally have a negative view about GAI replacing humans and no need of learning. They do not have negative views about GAI producing documents unless GAI is faulty. (4) Medical industry workers positively view GAI if it operates following their direction. However, they do not agree with the view that GAI replaces humans, and that human learning is unnecessary. (5) Females are less likely than men to have a positive view of GAI. In summary, views about GAI vary widely by the individual characteristics and condition of GAI, and by the question set.

document thumbnail

Par les mêmes auteurs

Sur les mêmes sujets

Sur les mêmes disciplines

Exporter en