8 links
tagged with all of: machine-learning + fine-tuning
Click any tag below to further narrow down your results
Links
LoRA has become a key method for fine-tuning large models, but its parameter redundancy limits efficiency. This research introduces SeLoRA, which employs spectral encoding to reduce redundancy without sacrificing expressiveness, demonstrating improved performance and efficiency across various tasks like commonsense reasoning and code generation.
LoRACode introduces a parameter-efficient fine-tuning method using Low-Rank Adaptation (LoRA) to improve code embeddings for semantic code search. The approach significantly reduces trainable parameters and enhances performance in code retrieval tasks, achieving notable gains in Mean Reciprocal Rank for both Code2Code and Text2Code searches across various programming languages. The authors provide their code and pre-trained models to support further research in this domain.
LoRA (Low-Rank Adaptation) is a parameter-efficient fine-tuning method that allows large language models to be updated with fewer parameters, making post-training faster and more resource-efficient. Recent experiments show that LoRA can achieve performance comparable to full fine-tuning (FullFT) under certain conditions, particularly with small-to-medium-sized datasets, but may struggle with larger datasets and high batch sizes. Key findings suggest a "low-regret regime" where LoRA's efficiency aligns with FullFT, paving the way for its broader application in various scenarios.
Large language models (LLMs) typically cannot adapt their weights dynamically to new tasks or knowledge. The Self-Adapting LLMs (SEAL) framework addresses this limitation by allowing models to generate their own finetuning data and directives for self-adaptation through a reinforcement learning approach, resulting in persistent weight updates and improved performance in knowledge incorporation and few-shot generalization tasks.
Fine-tuning large language models (LLMs) enhances their performance for specific tasks, making them more effective and aligned with user needs. The article discusses the importance of fine-tuning LLMs and provides a guide on how to get started, including selecting the right datasets and tools.
The article discusses the process of reinforcement learning fine-tuning, detailing how to enhance model performance through specific training techniques. It emphasizes the importance of tailored approaches to improve the adaptability and efficiency of models in various applications. The information is aimed at practitioners looking to leverage reinforcement learning for real-world tasks.
Fine-tuning a language model using LoRA (Low-Rank Adaptation) allows for efficient specialization without overwriting existing knowledge. The article details a hands-on experiment to adapt the Gemma 3 270M model for reliably masking personally identifiable information (PII) in text, showcasing the process of preparing a dataset, adding adapter layers, and training the model efficiently. Docker's ecosystem simplifies the entire fine-tuning workflow, making it accessible without requiring extensive resources.
Character.AI has open-sourced pipeling-sft, a scalable framework designed for fine-tuning large-scale MoE LLMs like DeepSeek V3. This framework addresses challenges in training efficiency and stability, integrating multi-level parallelism and supporting various precision formats, while facilitating seamless HuggingFace integration for researchers.