Envíos recentes

  • GRALENIA: Antimicrobial Resistance Management based on Natural Language and Artificial Intelligence 

    Bernardo-Castiñeira, Cristóbal; Bou, Germán; Campos, Manuel; Cánovas-Segura, Bernardo; Figueiras Gómez, Sergio; Gómez-Rodríguez, Carlos; Míguez-Rey, Enrique; Vilares, Jesús (CEUR-WS, 2024)
    [Abstract]: The objective of GRALENIA project is to develop a multidisciplinary, comprehensive and interoperable platform incorporating artificial intelligence algorithms and natural language processing techniques to improve ...
  • Dancing in the Syntax Forest: Fast, Accurate and Explainable Sentiment Analysis with SALSA 

    Gómez-Rodríguez, Carlos; Imran, Muhammad; Vilares, David; Solera, Elena; Kellert, Olga (CEUR-WS, 2024)
    [Abstract]: Sentiment analysis is a key technology for companies and institutions to gauge public opinion on products, services or events. However, for large-scale sentiment analysis to be accessible to entities with modest ...
  • Grammar Assistance Using Syntactic Structures (GAUSS) 

    Zamaraeva, Olga; Suárez Allegue, Lorena; Gómez-Rodríguez, Carlos; Alonso-Ramos, Margarita; Ogneva, Anastasiia (CEUR-WS, 2024)
    [Abstract]: Automatic grammar coaching serves an important purpose of advising on standard grammar varieties while not imposing social pressures or reinforcing established social roles. Such systems already exist but most ...
  • From Tokens to Trees: Mapping Syntactic Structures in the Deserts of Data-Scarce Languages 

    Vilares, David; Muñoz Ortiz, Alberto (CEUR-WS, 2024-06)
    [Abstract]: Low-resource learning in natural language processing focuses on developing effective resources, tools, and technologies for languages that are less popular within the industry and academia. This effort is crucial ...
  • BERTbek: A Pretrained Language Model for Uzbek 

    Kuriyozov, Elmurod; Vilares, David; Gómez-Rodríguez, Carlos (European Language Resources Association (ELRA), 2024-05)
    [Abstract]: Recent advances in neural networks based language representation made it possible for pretrained language models to outperform previous models in many downstream natural language processing (NLP) tasks. These ...

Máis