Skip to main content

Questions tagged [gpu]

-2 votes
3 answers
356 views

Backstory: Writing a QImage to Sixel renderer. Feel like I have optimized it the best I can using basic c++. I have heard suggestions here or there that you can utilize things like GPU, SIMD, insert ...
Anon's user avatar
  • 3,649
0 votes
1 answer
104 views

I was using the Standard_NC48ads_A100_v4 Compute cluster with 2 nodes to fine-tune a Phi-3-mini-128k-instruct model. The data used for the same was around 25 MB. I set the training batch size to 10. ...
S R's user avatar
  • 11
2 votes
0 answers
179 views

I am developing a desktop application that utilises Tensorflow. The aim of the application is to let users easily train a given model and use it for inference within the app. I want to support ...
turnip's user avatar
  • 1,701
4 votes
3 answers
1k views

My understanding is that every PRNG or QRNG requires a state to prevent the next item in its sequence from being too predictable; which is sensible, as they're all running on deterministic hardware. ...
Michael Macha's user avatar
2 votes
1 answer
213 views

Recently I ran against this question and this Nvidia-docker project, which is an Nvidia Docker implementation and it made me wondering where, why and how this scheme makes sense? I found out some ...
Suncatcher's user avatar
1 vote
2 answers
320 views

Basically, I am wondering what sort of speed I will get by parallelizing a algorithm to work with GPUs. I am wondering if someone has implemented queueing theory/Amdahl's law with a UI or if everyone ...
Robert Baron's user avatar
  • 1,132
7 votes
1 answer
5k views

I have come across the word "warp" in a few places but haven't seen a thorough definition (there's no Wikipedia page on it either). A brief definition is found here: In the SIMT paradigm, threads ...
Lance Pollard's user avatar
0 votes
1 answer
505 views

Trying to imagine how you would go about implementing summation (or reduction?) on a parallel architecture and am having a difficult time. Specifically thinking in terms of WebGL arrays of vectors ...
Lance Pollard's user avatar
4 votes
3 answers
3k views

I have seen a few papers on parallel/GPU processing of trees, but after briefly looking through them I wasn't able to grasp what they did. The closest to a helpful explanation was found in ...
Lance Pollard's user avatar
0 votes
1 answer
2k views

This is more of a Computer Engineering question, but what is the feature of a CPU to run Javascript fast? I use to access the internet with an AMD Phenom II with 6 cores and I could almost have as ...
Dehbop's user avatar
  • 169
7 votes
0 answers
224 views

There is an existing real-time, scientific visualization application that uses OpenCL and OpenGL to render complex 2D graphs. My goal is to incorporate this application into a 3D rendered scene. At ...
Liam Kelly's user avatar
2 votes
1 answer
227 views

I've spent the last few days working with tensorflow for the first time as part of a natural language processing assignment for my degree. It's been interesting (fun isn't the right word) trying to ...
HJCee's user avatar
  • 165
42 votes
7 answers
10k views

This is a general question on a subject I've found interesting as a gamer: CPU/GPU bottlenecks and programming. If I'm not mistaken, I've come to understand that both CPU and GPU calculate stuff, but ...
Azami's user avatar
  • 539
1 vote
1 answer
455 views

I have a perfectly parallel function that would run great on a machine with 1024 cores and 4GB RAM. There's quite a lot of branching (doing set union and traversing structs). There is no communication ...
Drathier's user avatar
  • 2,883
8 votes
1 answer
3k views

I am doing some research as to how most hardware accelerated GUI libraries work. I am actually only care about the rendering backends of them here. I am trying to figure out what would be the best way ...
Gerharddc's user avatar
  • 191

15 30 50 per page