Ilker Kesen
2022
CRAFT: A Benchmark for Causal Reasoning About Forces and inTeractions
Tayfun Ates
|
M. Ateşoğlu
|
Çağatay Yiğit
|
Ilker Kesen
|
Mert Kobas
|
Erkut Erdem
|
Aykut Erdem
|
Tilbe Goksun
|
Deniz Yuret
Findings of the Association for Computational Linguistics: ACL 2022
Humans are able to perceive, understand and reason about causal events. Developing models with similar physical and causal understanding capabilities is a long-standing goal of artificial intelligence. As a step towards this direction, we introduce CRAFT, a new video question answering dataset that requires causal reasoning about physical forces and object interactions. It contains 58K video and question pairs that are generated from 10K videos from 20 different virtual environments, containing various objects in motion that interact with each other and the scene. Two question categories in CRAFT include previously studied descriptive and counterfactual questions. Additionally, inspired by the Force Dynamics Theory in cognitive linguistics, we introduce a new causal question category that involves understanding the causal interactions between objects through notions like cause, enable, and prevent. Our results show that even though the questions in CRAFT are easy for humans, the tested baseline models, including existing state-of-the-art methods, do not yet deal with the challenges posed in our benchmark.
Detecting Euphemisms with Literal Descriptions and Visual Imagery
Ilker Kesen
|
Aykut Erdem
|
Erkut Erdem
|
Iacer Calixto
Proceedings of the 3rd Workshop on Figurative Language Processing (FLP)
This paper describes our two-stage system for the Euphemism Detection shared task hosted by the 3rd Workshop on Figurative Language Processing in conjunction with EMNLP 2022. Euphemisms tone down expressions about sensitive or unpleasant issues like addiction and death. The ambiguous nature of euphemistic words or expressions makes it challenging to detect their actual meaning within a context. In the first stage, we seek to mitigate this ambiguity by incorporating literal descriptions into input text prompts to our baseline model. It turns out that this kind of direct supervision yields remarkable performance improvement. In the second stage, we integrate visual supervision into our system using visual imageries, two sets of images generated by a text-to-image model by taking terms and descriptions as input. Our experiments demonstrate that visual supervision also gives a statistically significant performance boost. Our system achieved the second place with an F1 score of 87.2%, only about 0.9% worse than the best submission.
Search
Co-authors
- Erkut Erdem 2
- Aykut Erdem 2
- Tayfun Ates 1
- M. Ateşoğlu 1
- Çağatay Yiğit 1
- show all...