Skip to main content

All Questions

0 votes
0 answers
39 views

The issue of mask fragmentation during SAM2 tracking

I am currently working on object tracking. I use Moondream2 to identify objects in the scene, filter out duplicate bounding boxes, and then use SAM2 to track the objects. During the tracking process, ...
Limit 's user avatar
1 vote
0 answers
87 views

Running DeepSeek-V3 inference without GPU (on CPU only)

I am trying to run the DeepSeek-V3 model inference on a remote machine (SSH). This machine does not have any GPU, but has many CPU cores. 1rst method/ I try to run the model inference using the ...
The_Average_Engineer's user avatar
1 vote
1 answer
96 views

Why does my Llama 3.1 model act differently between AutoModelForCausalLM and LlamaForCausalLM?

I have one set of weights, one tokenizer, the same prompt, and identical generation parameters. Yet somehow, when I load the model using AutoModelForCausalLM, I get one output, and when I construct it ...
han mo's user avatar
  • 23
0 votes
0 answers
88 views

RuntimeError: Expected tensor for argument #1 'indices' to have one of the following scalar types: Long, Int; but got MPSFloatType instead

I want to apply Integrated Gradients as XAI method on a pretrained huggingface model. Inputs are movie reviews and I use Captum to generate attributions for each token. But I don't get my code to run. ...
human0id's user avatar
0 votes
0 answers
77 views

How can I group PyTorch Profiler events by layer hierarchy when profiling a Hugging Face Transformer?

I'm using PyTorch Profiler to inspect inference performance on a Hugging Face Transformer (e.g., Qwen model). I have code that successfully captures operator-level profiling information (like aten::mm,...
AlexL's user avatar
  • 1
2 votes
1 answer
89 views

"Inconsistent Predictions in PyTorch Model: Single Image vs. Batch Processing"

I am noticing a significant difference in model predictions when running predictions on a single image versus the whole dataset. The model, which was trained using PyTorch, gives drastically different ...
Dishant Dua's user avatar
0 votes
0 answers
62 views

Diffuser pipeline embedings not enough values to unpack

I wanted to generate a image using text embedding instead of text as input using clip to tokenizes & embeds. The code so far : from transformers import AutoTokenizer, CLIPTextModelWithProjection ...
Felox's user avatar
  • 502
0 votes
0 answers
80 views

How to use Inception V3 as Backbone for Vision Transformer?

I’m looking to create a Vision Transformer (ViT) using Inception V3 as the backbone. For an input image of size 500x500x3, Inception V3 outputs feature maps with dimensions [1, 2048, 14, 14]. How can ...
Asif Khan's user avatar
  • 1,278
0 votes
1 answer
77 views

An error occurs during the execution of UNet when the batch size is not equal to 1

I'm trying to run a Stable Diffusion model using the code provided in the DDIM Inversion tutorial. However, when the input's batch size is set to a value greater than 1 (e.g., 32), I encounter the ...
young's user avatar
  • 11
0 votes
0 answers
70 views

ValueError: If no `decoder_input_ids` or `decoder_inputs_embeds` are passed, `input_ids` cannot be `None`

I am trying to get the decoder hidden state of the florence 2 model. I was following this https://huggingface.co/microsoft/Florence-2-large/blob/main/modeling_florence2.py to understand the parameters ...
user10418143's user avatar
2 votes
1 answer
429 views

How does the data splitting actually work in Multi GPU Inference for Accelerate when used in a batched inference setting?

I followed the code given in this github issue and this medium blog I ran the batched experiment with process = 1 and process=4 it gave me the result but I'm confused right now because I thought the ...
Deshwal's user avatar
  • 4,232
0 votes
0 answers
67 views

How to create a custom model with Hugginface PreTrainedModel

I'm trying to create a simple model with the code below, taken almost directly from the documentation, and receiving an error import torch from transformers import PretrainedConfig, PreTrainedModel ...
Tobi's user avatar
  • 1
0 votes
1 answer
542 views

OutOfMemoryError: CUDA out of memory while using compute_metrics function in Hugging Face Trainer

I'm encountering a CUDA out of memory error when using the compute_metrics function with the Hugging Face Trainer during model evaluation. My GPU is running out of memory while trying to compute the ...
KainnT's user avatar
  • 15
0 votes
1 answer
84 views

Error using LlmFactory with "TheBloke/OpenHermes-2.5-Mistral-7B-GGUF" Huggingface

I tried replicating a simple Python code to create a small LLM model. I have macOS M1 machine. I created a separate environment where I installed Pytorch and llama-cpp-python. The code: from llmflex ...
Toly's user avatar
  • 3,189
1 vote
1 answer
3k views

Finding config.json for Llama 3.1 8B

I installed the Llama 3.1 8B model through Meta's Github page, but I can't get their example code to work. I'm running the following code in the same directory as the Meta-Llama-3.1-8B folder: import ...
MatthewScarpino's user avatar

15 30 50 per page
1
2 3 4 5 6