This project is a step-by-step learning journey where we implement various types of Triton kernels—from the simplest examples to more advanced applications—while exploring GPU programming with Triton.
On the 15th, local time, in a Stanford University Department of Computer Science (Computer Science) AI (Artificial ...
Abstract: Recent studies proposed to leverage large language models (LLMs) with In-Context Learning (ICL) to handle code intelligence tasks without fine-tuning. ICL employs task instructions and a set ...
Abstract: The Mixture of Experts (MoE) model is a promising approach for handling code-switching speech recognition (CS-ASR) tasks. However, the existing CS-ASR work on MoE has yet to leverage the ...
If you just want to use MIR as the pre-training indicator of your own model, no additional environment is required. python mir.py --model_path PATH/TO/MODEL --base_llm PATH/TO/LLM --text_data_path ...