Always private
DuckDuckGo never tracks your searches.
Learn More
You can hide this reminder in Search Settings
All regions
Argentina
Australia
Austria
Belgium (fr)
Belgium (nl)
Brazil
Bulgaria
Canada (en)
Canada (fr)
Catalonia
Chile
China
Colombia
Croatia
Czech Republic
Denmark
Estonia
Finland
France
Germany
Greece
Hong Kong
Hungary
Iceland
India (en)
Indonesia (en)
Ireland
Israel (en)
Italy
Japan
Korea
Latvia
Lithuania
Malaysia (en)
Mexico
Netherlands
New Zealand
Norway
Pakistan (en)
Peru
Philippines (en)
Poland
Portugal
Romania
Russia
Saudi Arabia
Singapore
Slovakia
Slovenia
South Africa
Spain (ca)
Spain (es)
Sweden
Switzerland (de)
Switzerland (fr)
Taiwan
Thailand (en)
Turkey
Ukraine
United Kingdom
US (English)
US (Spanish)
Vietnam (en)
Safe search: moderate
Strict
Moderate
Off
Any time
Any time
Past day
Past week
Past month
Past year
  1. aclanthology.org

    Jan 12, 2025%0 Conference Proceedings %T xDial-Eval: A Multilingual Open-Domain Dialogue Evaluation Benchmark %A Zhang, Chen %A D'Haro, Luis %A Tang, Chengguang %A Shi, Ke %A Tang, Guohua %A Li, Haizhou %Y Bouamor, Houda %Y Pino, Juan %Y Bali, Kalika %S Findings of the Association for Computational Linguistics: EMNLP 2023 %D 2023 %8 December %I ...
  2. Oct 13, 2023Recent advancements in reference-free learned metrics for open-domain dialogue evaluation have been driven by the progress in pre-trained language models and the availability of dialogue data with high-quality human annotations. However, current studies predominantly concentrate on English dialogues, and the generalization of these metrics to other languages has not been fully examined. This ...
  3. to the absence of a multilingual dialogue evalu-ation benchmark. To address the issue, we in-troduce xDial-Eval, built on top of open-source English dialogue evaluation datasets. xDial-Eval includes 12 turn-level and 6 dialogue-level English datasets, comprising 14930 an-notated turns and 8691 annotated dialogues respectively.
  4. ar5iv.labs.arxiv.org

    This paper introduces xDial-Eval, a multilingual dialogue evaluation benchmark featuring 14930 annotated turns and 8691 dialogues in 10 languages. Both automatic and human evaluation validate the high quality of xDial-Eval. Additionally, we examine the performance of BERT-based metrics and emerging LLMs on this benchmark.
  5. mendeley.com

    (2023) Zhang et al. Findings of the Association for Computational Linguistics: EMNLP 2023. Recent advancements in reference-free learned metrics for open-domain dialogue evaluation have been driven by the progress in pre-trained language models and the availability of dialogue data with high-qual...
  6. aclanthology.org

    This is largely due to the absence of a multilingual dialogue evaluation benchmark. To address the issue, we introduce xDial-Eval, built on top of open-source English dialogue evaluation datasets. xDial-Eval includes 12 turn-level and 6 dialogue-level English datasets, comprising 14930 annotated turns and 8691 annotated dialogues respectively.
  7. researchgate.net

    Jan 1, 2023Request PDF | On Jan 1, 2023, Chen Zhang and others published xDial-Eval: A Multilingual Open-Domain Dialogue Evaluation Benchmark | Find, read and cite all the research you need on ResearchGate
  8. gengo.sotaro.io

    Currently, human evaluation is the most reliable way to holistically judge the quality of the dialogue. Approach: They propose to use English dialogue evaluation metrics to generalize them to other languages. Outcome: The proposed metrics outperform OpenAI's ChatGPT in terms of average Pearson correlations over all datasets and languages.
  9. Bibliographic details on xDial-Eval: A Multilingual Open-Domain Dialogue Evaluation Benchmark. Stop the war! Остановите войну! solidarity - - ... A Multilingual Open-Domain Dialogue Evaluation Benchmark. CoRR abs/2310.08958 (2023) a service of . home. blog; statistics; update feed; XML dump; RDF dump; browse. persons; conferences;

    Can’t find what you’re looking for?

    Help us improve DuckDuckGo searches with your feedback

Custom date rangeX