scholarly journals Data-Intensive Experimental Linguistics

2011 ◽  
Vol 6 ◽  
Author(s):  
Steven Abney

Computational linguistics is not a specialization of linguistics at all; it is a branch of computer science. A large majority of computational linguists have degrees in computer science and positions in computer science departments. It was founded as an offshoot of an engineering discipline (machine translation), and has been subsequently shaped by its place within artificial intelligence, and by a heavy influx of theory and method from speech recognition (another engineering discipline) and machine learning. But computation is a means to an end; the essential feature is data collection, analysis, and prediction on the large scale. I will call it data-intensive experimental linguistics. I wish to explain how data-intensive linguistics differs from mainstream practice, why I consider it to be genuine linguistics, and why I believe that it enables fundamental advances in our understanding of language.

AI Magazine ◽  
2017 ◽  
Vol 38 (4) ◽  
pp. 99-106
Author(s):  
Jeannette Bohg ◽  
Xavier Boix ◽  
Nancy Chang ◽  
Elizabeth F. Churchill ◽  
Vivian Chu ◽  
...  

The Association for the Advancement of Artificial Intelligence, in cooperation with Stanford University's Department of Computer Science, presented the 2017 Spring Symposium Series, held Monday through Wednesday, March 27–29, 2017 on the campus of Stanford University. The eight symposia held were Artificial Intelligence for the Social Good (SS-17-01); Computational Construction Grammar and Natural Language Understanding (SS-17-02); Computational Context: Why It's Important, What It Means, and Can It Be Computed? (SS-17-03); Designing the User Experience of Machine Learning Systems (SS-17-04); Interactive Multisensory Object Perception for Embodied Agents (SS-17-05); Learning from Observation of Humans (SS-17-06); Science of Intelligence: Computational Principles of Natural and Artificial Intelligence (SS-17-07); and Wellbeing AI: From Machine Learning to Subjectivity Oriented Computing (SS-17-08). This report, compiled from organizers of the symposia, summarizes the research that took place.


Author(s):  
Francesco Di Tria

Ethics is a research field that is obtaining more and more attention in Computer Science due to the proliferation of artificial intelligence software, machine learning algorithms, robot agents (like chatbot), and so on. Indeed, ethics research has produced till now a set of guidelines, such as ethical codes, to be followed by people involved in Computer Science. However, a little effort has been spent for producing formal requirements to be included in the design process of software able to act ethically with users. In the paper, we investigate those issues that make a software product ethical and propose a set of metrics devoted to quantitatively evaluate if a software product can be considered ethical or not.


Author(s):  
P. Alison Paprica ◽  
Frank Sullivan ◽  
Yin Aphinyanaphongs ◽  
Garth Gibson

Many health systems and research institutes are interested in supplementing their traditional analyses of linked data with machine learning (ML) and other artificial intelligence (AI) methods and tools. However, the availability of individuals who have the required skills to develop and/or implement ML/AI is a constraint, as there is high demand for ML/AI talent in many sectors. The three organizations presenting are all actively involved in training and capacity building for ML/AI broadly, and each has a focus on, and/or discrete initiatives for, particular trainees. P. Alison Paprica, Vector Institute for artificial intelligence, Institute for Clinical Evaluative Sciences, University of Toronto, Canada. Alison is VP, Health Strategy and Partnerships at Vector, responsible for health strategy and also playing a lead role in “1000AIMs” – a Vector-led initiative in support of the Province of Ontario’s \$30 million investment to increase the number of AI-related master’s program graduates to 1,000 per year within five years. Frank Sullivan, University of St Andrews Scotland. Frank is a family physician and an associate director of HDRUK@Scotland. Health Data Research UK \url{https://hdruk.ac.uk/} has recently provided funding to six sites across the UK to address challenging healthcare issues through use of data science. A 50 PhD student Doctoral Training Scheme in AI has also been announced. Each site works in close partnership with National Health Service bodies and the public to translate research findings into benefits for patients and populations. Yin Aphinyanaphongs – INTREPID NYU clinical training program for incoming clinical fellows. Yin is the Director of the Clinical Informatics Training Program at NYU Langone Health. He is deeply interested in the intersection of computer science and health care and as a physician and a scientist, he has a unique perspective on how to train medical professionals for a data drive world. One version of this teaching process is demonstrated in the INTREPID clinical training program. Yin teaches clinicians to work with large scale data within the R environment and generate hypothesis and insights. The session will begin with three brief presentations followed by a facilitated session where all participants share their insights about the essential skills and competencies required for different kinds of ML/AI application and contributions. Live polling and voting will be used at the end of the session to capture participants’ view on the key learnings and take away points. The intended outputs and outcomes of the session are: Participants will have a better understanding of the skills and competencies required for individuals to contribute to AI applications in health in various ways Participants will gain knowledge about different options for capacity building from targeted enhancement of the skills of clinical fellows, to producing large number of applied master’s graduates, to doctoral-level training After the session, the co-leads will work together to create a resource that summarizes the learnings from the session and make them public (though publication in a peer-reviewed journal and/or through the IPDLN website)


Author(s):  
Yung Ming ◽  
Lily Yuan

Machine Learning (ML) and Artificial Intelligence (AI) methods are transforming many commercial and academic areas, including feature extraction, autonomous driving, computational linguistics, and voice recognition. These new technologies are now having a significant effect in radiography, forensics, and many other areas where the accessibility of automated systems may improve the precision and repeatability of essential job performance. In this systematic review, we begin by providing a short overview of the different methods that are currently being developed, with a particular emphasis on those utilized in biomedical studies.


Arts ◽  
2019 ◽  
Vol 8 (1) ◽  
pp. 26 ◽  
Author(s):  
Marian Mazzone ◽  
Ahmed Elgammal

Our essay discusses an AI process developed for making art (AICAN), and the issues AI creativity raises for understanding art and artists in the 21st century. Backed by our training in computer science (Elgammal) and art history (Mazzone), we argue for the consideration of AICAN’s works as art, relate AICAN works to the contemporary art context, and urge a reconsideration of how we might define human and machine creativity. Our work in developing AI processes for art making, style analysis, and detecting large-scale style patterns in art history has led us to carefully consider the history and dynamics of human art-making and to examine how those patterns can be modeled and taught to the machine. We advocate for a connection between machine creativity and art broadly defined as parallel to but not in conflict with human artists and their emotional and social intentions of art making. Rather, we urge a partnership between human and machine creativity when called for, seeing in this collaboration a means to maximize both partners’ creative strengths.


2021 ◽  
Vol 5 (2) ◽  
pp. 113
Author(s):  
Youngseok Lee ◽  
Jungwon Cho

In the near future, as artificial intelligence and computing network technology develop, collaboration with artificial intelligence (AI) will become important. In an AI society, the ability to communicate and collaborate among people is an important element of talent. To do this, it is necessary to understand how artificial intelligence based on computer science works. AI is being rapidly applied across industries and is developing as a core technology to enable a society led by knowledge and information. An AI education focused on problem solving and learning is efficient for computer science education. Thus, the time has come to prepare for AI education along with existing software education so that they can adapt to the social and job changes enabled by AI. In this paper, we explain a classification method for AI machine learning models and propose an AI education model using teachable machines. Non-computer majors can understand the importance of data and the AI model concept based on specific cases using AI education tools to understand and experiment with AI even without the knowledge of mathematics, and use languages such as Python, if necessary. Through the application of the machine learning model, AI can be smoothly utilized in their field of interest. If such an AI education model is activated, it will be possible to suggest the direction of AI education for collaboration with AI experts through the application of AI technology.


2022 ◽  
Vol 14 (2) ◽  
pp. 1-15
Author(s):  
Lara Mauri ◽  
Ernesto Damiani

Large-scale adoption of Artificial Intelligence and Machine Learning (AI-ML) models fed by heterogeneous, possibly untrustworthy data sources has spurred interest in estimating degradation of such models due to spurious, adversarial, or low-quality data assets. We propose a quantitative estimate of the severity of classifiers’ training set degradation: an index expressing the deformation of the convex hulls of the classes computed on a held-out dataset generated via an unsupervised technique. We show that our index is computationally light, can be calculated incrementally and complements well existing ML data assets’ quality measures. As an experimentation, we present the computation of our index on a benchmark convolutional image classifier.


2020 ◽  
Vol 245 ◽  
pp. 08020
Author(s):  
Ian Bird ◽  
Simone Campana ◽  
Pere Mato Vila ◽  
Stefan Roiser ◽  
Markus Schulz ◽  
...  

With the ever-increasing size of scientific collaborations and complexity of scientific instruments, the software needed to acquire, process and analyze the gathered data is increasing in both complexity and size. Unfortunately the role and career path of scientists and engineers working on software R&D and developing scientific software are neither clearly established nor defined in many fields of natural science. In addition, the exchange of information between scientific software development and computer science departments at universities or computing schools is scattered and fragmented into individual initiatives. To address the above issues we propose a new effort on a European level, which concentrates on strengthening the role of software developers in natural sciences, acts as a hub for the exchange of ideas among different stake-holders in computer science and scientific software and forms a lobbying forum for software engineering in natural sciences on an international level. This contribution discusses in detail the motivation, role and interplay with other initiatives of a “Software Institute for Data-Intensive Sciences”, which is currently being discussed between research institutes, universities and funding agencies in Europe. In addition to the current status, an outlook on future prospects of this initiative will be given.


Author(s):  
M. Stashevskaya

The article contains a study of existing views on the economic content of big data. From among the views, within which the authors define big data, the descriptive-model, utility-digital and complex-technological approaches are formulated. Against the back- ground of the large-scale spread of digital technologies (machine learning, cloud computing, artificial intelligence, augmented and virtual reality, etc.), functioning thanks to big data, the study of their economic essence is becoming especially relevant. As a result, it was found that the basis of economic activity in the digital economy is big data. The definition of big data as a resource of the digital economy is proposed.


Sign in / Sign up

Export Citation Format

Share Document