Skip to main content
Stack Overflow
  1. About
  2. For Teams
Filter by
Sorted by
Tagged with
1 vote
0 answers
69 views

My proxy goal is to change LoRA from h = (W +BA)x to h = (W + BAP)x. Preliminary code attached for your reference My actual goal is to train a model with the following loss: 〖Θ ̃=(arg min)┬Δ ̂ 〗⁡〖‖f_(...
3 votes
0 answers
63 views

I am trying to deploy a fine-tuned Mistral-7B model on an Azure ML Online Endpoint. The deployment repeatedly fails during the init() phase of the scoring script with an huggingface_hub.errors....
1 vote
0 answers
66 views

I'm fine-tuning a CrossEncoder model with LoRA using sentence-transformers library on Kaggle (12-hour limit). I need to resume training from a checkpoint, but I'm getting a ValueError when trying to ...
1 vote
0 answers
119 views

I’ve been working on fine-tuning LLaMA 2–7B using QLoRA with bitsandbytes 4-bit quantization and ran into a weird issue. I did adaptive pretraining on Arabic data with a custom tokenizer (vocab size ~...
0 votes
1 answer
345 views

I would like to perform continued pretraining of Qwen2.5-VL-Instruct using Unsloth + LoRA, following a three-stage training process: Stage 1: Train only the projector (Alignment) Stage 2: Train both ...
2 votes
1 answer
815 views

I'm using Unsloth's FastVisionModel with the base model unsloth/qwen2-VL-2B-Instruct to train on a dataset that includes text with many unique characters. Here's the overall process I followed: ...
0 votes
0 answers
156 views

I'm implementing a more efficient version of lokr.Linear from the LoKr module in PEFT. The current implementation uses torch.kron to construct the delta_weight before applying rank dropout, but this ...
0 votes
1 answer
635 views

I'm facing an issue when training a model using PEFT and LoRA on a multi-GPU setup with PyTorch and Hugging Face Transformers. The error I get is: RuntimeError: Expected all tensors to be on the same ...
9 votes
1 answer
6k views

Question: I am trying to fine-tune the Mistral-7B-Instruct-v0.1-GPTQ model using SFTTrainer from trl. However, when running my script in Google Colab, I encounter the following error: TypeError: ...
0 votes
0 answers
219 views

I'm having trouble loading my LoRA adapters for inference after fine-tuning Llama 3.1 8B. When I try to load the adapter files in a new session, I get a warning about missing adapter keys: /usr/local/...
3 votes
1 answer
2k views

I am trying to fine-tune a language model using SFTTrainer from the trl library in Google Colab. However, I am encountering the following error: TypeError Traceback (...
1 vote
0 answers
85 views

We all know that LoRA is a low-rank adaptation method, which can be formulated as follows: x = W_0 * x + (A @ B) * x. I have two different code implementations of this. Are there any differences ...
1 vote
1 answer
172 views

I tried to merge a pefted model to the original one. Cause Hugging face APi only outputs "extra weights" of the fine-tuning as a .safetensors file. I try to merge but failed. I wonder how ...
1 vote
1 answer
292 views

I am trying to finetune the nvidia/NV-Embed-v2 model from hugging face using lora from peft library. I am facing the "TypeError: NVEmbedModel.forward() got an unexpected keyword argument '...
2 votes
0 answers
648 views

I am trying to inference a model ChemVLM (https://huggingface.co/AI4Chem/ChemVLM-26B). While trying to run the python code, I get error ImportError: This modeling file requires the following packages ...

15 30 50 per page
1
2 3 4 5

AltStyle によって変換されたページ (->オリジナル) /