All Questions
87 questions
0
votes
0
answers
39
views
The issue of mask fragmentation during SAM2 tracking
I am currently working on object tracking. I use Moondream2 to identify objects in the scene, filter out duplicate bounding boxes, and then use SAM2 to track the objects. During the tracking process, ...
1
vote
0
answers
87
views
Running DeepSeek-V3 inference without GPU (on CPU only)
I am trying to run the DeepSeek-V3 model inference on a remote machine (SSH). This machine does not have any GPU, but has many CPU cores.
1rst method/
I try to run the model inference using the ...
1
vote
1
answer
96
views
Why does my Llama 3.1 model act differently between AutoModelForCausalLM and LlamaForCausalLM?
I have one set of weights, one tokenizer, the same prompt, and identical generation parameters. Yet somehow, when I load the model using AutoModelForCausalLM, I get one output, and when I construct it ...
0
votes
0
answers
88
views
RuntimeError: Expected tensor for argument #1 'indices' to have one of the following scalar types: Long, Int; but got MPSFloatType instead
I want to apply Integrated Gradients as XAI method on a pretrained huggingface model. Inputs are movie reviews and I use Captum to generate attributions for each token.
But I don't get my code to run. ...
0
votes
0
answers
77
views
How can I group PyTorch Profiler events by layer hierarchy when profiling a Hugging Face Transformer?
I'm using PyTorch Profiler to inspect inference performance on a Hugging Face Transformer (e.g., Qwen model). I have code that successfully captures operator-level profiling information (like aten::mm,...
2
votes
1
answer
89
views
"Inconsistent Predictions in PyTorch Model: Single Image vs. Batch Processing"
I am noticing a significant difference in model predictions when running predictions on a single image versus the whole dataset. The model, which was trained using PyTorch, gives drastically different ...
0
votes
0
answers
62
views
Diffuser pipeline embedings not enough values to unpack
I wanted to generate a image using text embedding instead of text as input using clip to tokenizes & embeds.
The code so far :
from transformers import AutoTokenizer, CLIPTextModelWithProjection
...
0
votes
0
answers
80
views
How to use Inception V3 as Backbone for Vision Transformer?
I’m looking to create a Vision Transformer (ViT) using Inception V3 as the backbone. For an input image of size 500x500x3, Inception V3 outputs feature maps with dimensions [1, 2048, 14, 14].
How can ...
0
votes
1
answer
77
views
An error occurs during the execution of UNet when the batch size is not equal to 1
I'm trying to run a Stable Diffusion model using the code provided in the DDIM Inversion tutorial. However, when the input's batch size is set to a value greater than 1 (e.g., 32), I encounter the ...
0
votes
0
answers
70
views
ValueError: If no `decoder_input_ids` or `decoder_inputs_embeds` are passed, `input_ids` cannot be `None`
I am trying to get the decoder hidden state of the florence 2 model. I was following this https://huggingface.co/microsoft/Florence-2-large/blob/main/modeling_florence2.py to understand the parameters ...
2
votes
1
answer
429
views
How does the data splitting actually work in Multi GPU Inference for Accelerate when used in a batched inference setting?
I followed the code given in this github issue and this medium blog
I ran the batched experiment with process = 1 and process=4 it gave me the result but I'm confused right now because I thought the ...
0
votes
0
answers
67
views
How to create a custom model with Hugginface PreTrainedModel
I'm trying to create a simple model with the code below, taken almost directly from the documentation, and receiving an error
import torch
from transformers import PretrainedConfig, PreTrainedModel
...
0
votes
1
answer
542
views
OutOfMemoryError: CUDA out of memory while using compute_metrics function in Hugging Face Trainer
I'm encountering a CUDA out of memory error when using the compute_metrics function with the Hugging Face Trainer during model evaluation. My GPU is running out of memory while trying to compute the ...
0
votes
1
answer
84
views
Error using LlmFactory with "TheBloke/OpenHermes-2.5-Mistral-7B-GGUF" Huggingface
I tried replicating a simple Python code to create a small LLM model.
I have macOS M1 machine.
I created a separate environment where I installed Pytorch and llama-cpp-python. The code:
from llmflex ...
1
vote
1
answer
3k
views
Finding config.json for Llama 3.1 8B
I installed the Llama 3.1 8B model through Meta's Github page, but I can't get their example code to work. I'm running the following code in the same directory as the Meta-Llama-3.1-8B folder:
import ...