We note that our work focuses on architectural comparisons rather than competing with recent SLM developments (e.g., SmolLM, MobileLLM). Our analysis isolates the fundamental advantages of ...
Large language models (LLMs) have changed the game for machine translation (MT). LLMs vary in architecture, ranging from decoder-only designs to encoder-decoder frameworks. Encoder-decoder models, ...
An Encoder-Decoder model is a fundamental architecture in the field of deep learning and natural language processing (NLP). It's widely used for a variety of tasks, including machine translation, text ...
Abstract: The main purpose of multimodal machine translation (MMT) is to improve the quality of translation results by taking the corresponding visual context as an additional input. Recently many ...
Abstract: Travel route recommendation is an important part of electronic tour guides and map applications. It aims to recommend a sequence of points of interest (POIs) to users based on their ...
Motivated by the success of T5 (Text-To-Text Transfer Transformer) in pre-trained natural language processing models, we propose a unified-modal SpeechT5 framework that explores the encoder-decoder ...
Computational modeling is an essential component of modern drug discovery. One of its most important applications is to select promising drug candidates for pharmacologically relevant target proteins.