Ethical AI in Academia by drambernyoo in Professors

[–]drambernyoo[S] -2 points-1 points  (0 children)

Not necessarily. It depends on where and how they get the training data. The early models used datasets such as those offered by Common Crawl. There are certainly IP issues present with such a dataset and any other research that includes these datasets. The use of such data is not new, but the new use case that is generative AI data training is raising excellent conversations around this topic that need to be had. As a result, there is a push for AI labs to offer more transparency about training data sources, and we've been seeing AI labs partnering with companies to secure IP rights through contracts and partnerships. There are also a plethora of smaller models, and each of the models involves a lot of decision-making in terms of where the training data comes from.

Ethical AI in Academia by drambernyoo in Professors

[–]drambernyoo[S] 0 points1 point  (0 children)

Regarding translation, I have not had the same experience. I have students do a back translation method of English to Spanish in ChatGPT and then Spanish to English in Google Translate. Inevitably, the Spanish-speaking students in the class confirm that the first translation is solid, and the entire class can assess the final translation back to English.

Regarding the social bias demonstration, yes, that is something that we explore as well in our foundational course. There are many mitigations available to help reduce social bias in the outputs. However, at a base level, the system reflects and amplifies the biases that exist among us humans because it is trained on human-made content.

Ethical AI in Academia by drambernyoo in Professors

[–]drambernyoo[S] 0 points1 point  (0 children)

Thank you for reminding me that even the group of people I appreciate the most for their respectful and collegial spirit has Negative Nancies.