Recent Submissions

  • GRALENIA: Antimicrobial Resistance Management based on Natural Language and Artificial Intelligence 

    Bernardo-Castiñeira, Cristóbal; Bou, Germán; Campos, Manuel; Cánovas-Segura, Bernardo; Figueiras Gómez, Sergio; Gómez-Rodríguez, Carlos; Míguez-Rey, Enrique; Vilares, Jesús (CEUR-WS, 2024)
    [Abstract]: The objective of GRALENIA project is to develop a multidisciplinary, comprehensive and interoperable platform incorporating artificial intelligence algorithms and natural language processing techniques to improve ...
  • Dancing in the Syntax Forest: Fast, Accurate and Explainable Sentiment Analysis with SALSA 

    Gómez-Rodríguez, Carlos; Imran, Muhammad; Vilares, David; Solera, Elena; Kellert, Olga (CEUR-WS, 2024)
    [Abstract]: Sentiment analysis is a key technology for companies and institutions to gauge public opinion on products, services or events. However, for large-scale sentiment analysis to be accessible to entities with modest ...
  • Grammar Assistance Using Syntactic Structures (GAUSS) 

    Zamaraeva, Olga; Suárez Allegue, Lorena; Gómez-Rodríguez, Carlos; Alonso-Ramos, Margarita; Ogneva, Anastasiia (CEUR-WS, 2024)
    [Abstract]: Automatic grammar coaching serves an important purpose of advising on standard grammar varieties while not imposing social pressures or reinforcing established social roles. Such systems already exist but most ...
  • From Tokens to Trees: Mapping Syntactic Structures in the Deserts of Data-Scarce Languages 

    Vilares, David; Muñoz Ortiz, Alberto (CEUR-WS, 2024-06)
    [Abstract]: Low-resource learning in natural language processing focuses on developing effective resources, tools, and technologies for languages that are less popular within the industry and academia. This effort is crucial ...
  • BERTbek: A Pretrained Language Model for Uzbek 

    Kuriyozov, Elmurod; Vilares, David; Gómez-Rodríguez, Carlos (European Language Resources Association (ELRA), 2024-05)
    [Abstract]: Recent advances in neural networks based language representation made it possible for pretrained language models to outperform previous models in many downstream natural language processing (NLP) tasks. These ...

View more