No.2270
.flip it
No.2273
I might buy a used 4090 since everyone's dumping them on the secondary market locally and the 5090 is shit.
I'll only give the envious green AI snake oil company money indirectly at this point.
No.2274
why are you buying an nvidia card. nvidia is an AI for buisness company
No.2275
Couldn't you get like two 3090s or two 4090s (or some other combo) used for the price of a new 5090? that would probably be better for AI than a single 5090. Or maybe a quadro and the likes. I wouldn't trust the 50 series cards just on principle that they lied about the performance metrics of the 5070, saying it's better than the 4090 at half the price.
No.2277
>>2276There's a kissu pc?
No.2278
>>2277Yeah it's what the Palworld server is hosted on
No.2279
I believe there are ways that it's used in LAN. The latency would probably be pretty bad by network
No.2280
Really it's about rendering rather than as a studio. Rendering for 3d is only ever done by distributed systems
No.2281
Pathetic
No.2282
>>2275Two GPUs unfortunately have a hidden cost to them. I'd need a new motherboard and if I need a new motherboard then I need a new CPU. I might already need a new PSU if I'm getting a 5090, so the costs are absurd. Bleh!
>>2276Not really, or at least I doubt it. That kind of setup is for rendering farms, once you have a complete scene and you send the data over to be processed, like a Pixar movie. The performance I'd want is for rotating the camera around without stuttering in a complex scene while having other 3D programs open or even a game window. You know, "live" stuff. Multi monitor setups are less impressive when you can't make use of them for everything.
No.2285
i don't even know if multi-gpu is a thing that works. when I tried to do it for video games it always had issues and the data had to be mirrored across both cards.
No.2286
>>2285Don't know about gaming, but that's how they do it at the big AI training servers with 100*s of GBs of VRAM.
No.2309
The 5090 doesn't seem dramatically better compared to the 4090, like the 4090 was compared to the 3090/Ti. It's mostly just a generational improvement. Real world performance seems to be about 30% higher. That performance increase seems basically identical to the increase in TDP; 450W on the 4090 compared to 575W on the 5090, or ~28% higher. The efficiency in terms of FPS per watt more less bears this out, with the 5090 being around the efficiency of the 4090, or slightly below.
In Gamers Nexus' testing (
https://www.youtube.com/watch?v=VWSlOC_jiLQ), for reference, the 5090 had an FPS/W of 0.34, and the 4090 at 0.35. The 5090 drew 538.4W, and the 4090 drew 391.7W.
Not exactly very impressive. It basically looks like the 50-series is going to be a refresh generation.
No.2310
>>2309What about for AI, like most people have been touting it for?
No.2311
>>2310Hard to say. AI compute could be anywhere from 1.3x to 2.5x greater than the 4090. You'll likely have to wait until the release date on January 30th when regular people get their hands on one.
If you believe Nvidia's marketing, it should be 2.5x the speed of the 4090, going by their reported Tensor core TOPS. But... Well... They also said the 5090 would be twice the performance of the 4090, which it very obviously isn't (
>>2309)...
No.2313
>>2311Yeah, I'm on at 3080 with 12gb of VRAM at the moment. It's a massive upgrade for AI just based on the VRAM alone. For a lot of the AI stuff you need a certain threshold of VRAM just to run it, like the video stuff I've mentioned before would take me 30 minutes to do what a 4090 does in a minute, and that's with making a lot of sacrifices. You can either store it in speedy VRAM or you can't.
>January 30thNice, I still have some time to think it over.
No.2316
>>2315Nice, thanks. The two yotubers I look at once in a while (gaming nexus and jayz2cents) just looked at the gaming side of things. I'm generally more interested in image/video AI more than text since text has such absurd VRAM requirements for the better models. I'm sure the efficiency there is increasing over time, or at least I hope.
I'm leaning towards getting the card, but I just learned that it's a PCIE 5 thing instead of the 4 my motherboard has. 5 is backwards compatible, but I won't get the best possible performance. I guess I could buy the card and get a new motherboard in summer or so.
No reports of its performance in 3D programs, though. I'm sure it's good, but I'd like to hear more about it.
No.2357
oh wait, now I see that I said something similar 2 days ago
being sick is fun
No.2412
>>2273Update: I think a lot of people had the same reaction to Blackwell being shit, all the local 4090's disappeared just a little after the official reviews broke.
No.2479
>>2478Yeah, he's not the first to call it a paper launch. I'm not even sure what my plans are now.
No.2480
>>2479if you've got 1000 dollars to burn buy a vr set and get a better idea of what you can do with 3D models+inspiration.
No.2481
>>2480VR is something I looked into for easy money once, but it's not something I'd want to deal with. The first thing is that it's Unity and I don't want to learn an engine whose primary focus is phone games. The second is that I'm not generally good at talking to new people and making ERP avatars for people is
very personal.
Has anyone said anything about more 5090s coming in at any point? Quarter 2 or summer or anything?
No.2483
>>2481also Unity is the engine with that guy that wanted to monetize it further and everyone rallied against online. If possible people should avoid Unity since he's just biding his time.
No.2486
>>2483sounds like a minor issue
No.2487
>>2486It really isn't a minor issue that a game engine seeks percentages of all your profits and wants to attempt to retroactively change things. It's a massive liability.
No.2488
>>2487it's minor because there's UE, or directly interacting with the XR APIs with software.
https://vroid.com/en/studio is where models are uploaded and "supports Web (API), Unity, and Unreal Engine."
No.2499
>>2498I mean, pretty much everyone and their mother was waiting for the 50 series to come out before upgrading...
No.2502
There's a table for comparing MHZ and cas latency. There was some other things about the number of chips on them or something.idk.
For boards... I think transfer speed is pretty fixed and it's down to stuff like ddr and pcie versions. And picking the best of a few brands..
No.2503
CPU SECURED! (I hope)
If somehow anyone else has been refreshing sites hoping to get the 9800X3D, it's on amazon! And it's directly available from amazon instead of mysterious chinese shops created days earlier.
https://www.amazon.com/dp/B0DKFMSMYK
No.2544
RAM is also here.
I'm going to bump to let people know this info here about 5090s likely being in high supply in a month:
>>2532
No.2545
>>2532Mmmmmm, seeing some conflicting info about the melting wires here. Is it something to be worried about on any setup, or are there easy ways to avoid it?
No.2546
>>2545There's also a thing about some cards not having sufficient ROPs and I don't know what a ROP is exactly but it means weaker performance:
https://videocardz.com/newz/nvidia-confirms-0-5-of-rtx-5090d-and-rtx-5070-ti-were-shipped-with-fewer-ropsIt's something you could check yourself and they'd replace your card for it, though.
If the news is true and they're pivoting to making lots of 5090s then I could see them looking more closely at the manufacturing process. The wires thing seems pretty rare, just like the 4090s, but who knows.
No.2548
>>2544this meower just rammed me
No.2576
File:mb.jpg (838.07 KB,1867x1249)

motherboard
No.2578
File:0.jpg (1.37 MB,1925x1323)

The case is quite nice, and it better be at the price. ($180, but I justified it because I buy a new one once a decade).
Lots of nice slots and it has a little plastic case for screws. It has a fan controller chipset thingie which is very helpful, I used to have one I bought separately and fit in a CD tray. Cases are quite roomy now without optical drives, huh.
No.2579
File:1.jpg (1.28 MB,1928x1336)

Replaced CPU AIO's fans since they're said to be loud. Got a 5 pack of thermaltake ones for $30.
No.2580
File:2.jpg (441.86 KB,1008x756)

Ram in motherboard, motherboard in case.
The motherboard is known as one of the cheapest X870 boards (ASRock X870 Pro RS). It performs well, but and they really skimped out on everything other than the motherboard. You don't even get a booklet. You get ONE SATA cable which made me laugh. I have to look online as to which slots are what, like the M2 slots. Well, I guess it's a good way to cut costs.
No.2602
>>2600oh and I did fix the Thermaltake logo thing, the top part pops off.
Ran Prime95 for an hour and CPU and memory didn't crash at all.
No.2606
The CPU cooler is using an "All in one" system as mentioned here
>>2543It's water cooling, but it's a closed system for one part. It's sealed up in the factory so I can't do anything with it. The benefit is there's really no chance of leaking, but the con is that it will very slowly evaporate over time and I can't replace the water. That should take years, though, and if it doesn't then at least it was the same price as an air cooler. (This one is highly rated)
Part of the reason I did it is that the 5090 will be massive and expel hot air towards CPU, so having a radiator at the top of the case for cooling will greatly benefit the CPU.
No.2636
>>2624That's pretty cool. It uses a vacuum or something? Back when I was doing my hobbyist watercooling stuff back in the 00s you would spend a few days running the loop with a separate PSU to check for leaks, but leaking almost never seemed dangerous to the computer since gravity naturally dropped the stuff away from parts.
The failure points on this AIO
>>2600 is where the tubes connect to the radiator in the top right (out of picture) and from the CPU block (the upside-down 'thermaltake' square) and in both cases the water would be falling straight down to the bottom of the case. GPU blocks similarly would be on the bottom of the card.
Water and electronics are definitely a risky combination, but it's not as bad as people think.
No.2638
>>2636>That's pretty cool. It uses a vacuum or something?If I remember right, it uses two separate pumps (one on the inlet and one on the outlet) and when the resistance falls due to a leak occurring, one of the pumps reverses flow so that the liquid is pulled into the reservoir.
No.2645
>>2638I was thinking it probably doesn't help well with punctures until a certain amount of liquid has spilled
No.2646
>>2645Yeah, I think it mainly makes sense as a prevention against server techs killing your $150,000 AI server (and every server below it) if they accidentally dislodge a tube off a barb fitting or something.
No.2649
File:m2.jpg (462.68 KB,1797x1257)

For some reason the new motherboard isn't reading my oldest M2 drive which is what had my profiles on it for browsers and stuff.
Bought a $20 external M2 SSD reader thingie and... thankfully the drive is working. It makes me wonder what is wrong with the motherboard, though, maybe the M2 is too old? I moved it to different slots and even tried removing another one to see if somehow it was a bandwidth thing or something.
Kind of annoying, but at least it's a small drive and I could put some less accessible stuff on it like older AI models or something.
No.2820
>>2819>$5500there are places where you could buy an apartment for the price of a few of these
No.2822
>>2820That price is actually pretty close what i bought my previous apartment for. Then again, that was like a decade ago and it's probably several times more expensive now, so i guess your point still stands.
Imagine putting "new graphics card" as the reason for a mortgage application...
No.3894
>>3893And yes, it has Ice Storm AND Fire Storm. I don't know how they don't cancel each other out, but that's the power of modern GPUs.
No.3896
>>3893>Sadly, my wonderful and beautiful computer parts model is no longer of this world. It puts things in perspective that time is limited.;_;
No.3927
>>3926That's a lot of Kuon, almost too much. Your Non-Fungible-Kuon stock is plummeting in value!
No.3929
>>3928but it really does bother me how it doesn't get her tail and ears correct all of the time!
No.3930
>>3928>Imagine endless randomized Kuon outfits in endlessly randomized accessories and randomized this and randomized expressions and randomized that and wow!you'll love the Utawarerumono gacha
No.3931
>>3930I am somehow familiar with it since I 'follow' the art uploads on an exhentai gallery. I'm not a fan of how it spits all over the story with time travel and other dimensions and the other gacha justifications to commodify the emotional connection to beloved characters.
I ignore everything about it apart from the nice art.
No.3932
>>3930>he fell for the gacha blunderNever spend money on gacha when the art is free and you can make your own combat in your head!
No.3937
>>2269>I'm in that CG tracker that went private 5 years ago so software and assets are no issuecgpeers? Can you see if something I've been looking for is there?
No.3938
I'm currently in the process of upgrading to the best hardware for editing + AI that my budget allows, which would be a 3090 Ti. I'm not too worried about not getting the 5090, but what hurts me is that I'm not able to afford a 4090, because the leap from the 30 series to it seems to be very significant.
I usually only do a full upgrade every decade or so, so I'm checking for motherboards with two PCIe 5.0 slots just in case I might be able to afford one or two capable cards for cheap in the future. I've found some 690 / 790 MBs that I might be able to get for around 150 bucks, if I get lucky.
For the CPU, I'm currently looking at the i9-12900KS, with the possibility to swap it out for a 14th gen in the future. It seems that the 14900KS will be the last for this socket, so I limited my RAM options to its 192 GB and DDR5 5600 maximum, even if the 12th gen I'm aiming for right now can't make use of it.
I'm still somewhat new to researching hardware and only ever got random pre-builds in the past, so I'm open to any criticism if something I said is actually nonsensical and I'm unaware of it. I tried my very hardest to understand how the hardware might interact with each other and to take it into account for possible seperate upgrades a few years down the line.
No.3939
i boughted a 4090
No.3940
>>3935One of the two classic blunders that one must know, the second, lesser known is this:
>Never go against a green texter when funposting is on the line.
No.4109
I'm not very good with hardware compared to most enthusiasts and was hoping /maho/ could please give me some help with something of a new build. I've been offered a free PC upgrade through my mother's company but I'm not confident in what I'm doing. From what I can read at the moment, AMD has overtaken Nvidia for practical affordability and price:performance ratio, but doesn't offer very high end GPUs. I've also been recommended an X3D CPU by an acquaintance, but I don't believe Intel offers an equivilent yet.
I've been using Intel+Nvidia all my life just because it's what I had help with from my older cousin growing up. My main use case is games, as well as some Adobe suite work mostly Lightroom and Photoshop and a small amount of video editing. I don't do anything in the vein of AI, and don't really play many new release games; my current hardware is able to run most of what I play without issue with the most demanding thing being Baldur's Gate 3, heavily modded Skyrim, or Team Fortress 2. Although TF2 is an older game it struggles on my current hardware and I believe simply upgrading the CPU will do wonders for it. I intend to play TES6 when it eventually releases.
https://nz.pcpartpicker.com/list/vgZ9LcI am already using the case, CPU fan and RAM in the above link. From what I can understand the 9800X3D and 9950X3D are functionally the same for my use case with the latter just being better for workstation use. According to this website and this review
https://www.videocardbenchmark.net/video_lookup.php?gpu=Radeon+RX+9070+XT&id=5956https://gamersnexus.net/gpus/amd-radeon-rx-9070-xt-gpu-review-benchmarks-vs-5070-ti-5070-7900-xt-sapphire-pulseThe 5070 Ti is slightly better than the 9070 XT at a slightly higher price point, but is overpriced compared to RRP.
One of the difficulties in doing this is that everything I get has to be from a local store called PBTech due to use of a business account to recieve a discount, so something like the 7900 XT is not an option. I'd like to please ask for help choosing a graphics card, motherboard and PSU, then RAM if necessary.
No.4110
>>3937>cgpeers? Can you see if something I've been looking for is there?Sorry I didn't see that post until now because someone bumped the thread. kissu has been too fast and active!
Yes I can, what are you looking for, person two weeks in the past?
>>3938VRAM is the biggest thing for AI. Sure, the 5090 and 4090 will have some speed over the 3090 but having the 24gb of VRAM on the 3090 is what allows you to do many things at all. A 3090 is like having a $400,000 Ferrari instead of a $500,000 Lamborghini. It's not an exaggeration that the change from a model being offloaded some to RAM from VRAM is like 30-50x slower, if it's possible at all.
>so I'm checking for motherboards with two PCIe 5.0 slotsBe aware that the newer cards are huge. I can't physically fit my old card in the motherboard on the second slot. I wonder if someone will make an aftermarket cooler that's thin, but I bet it will be expensive as all of these things are.
>For the CPU, I'm currently looking at the i9-12900KS,When I did research for CPU stuff AMD was the kind of mixed work/gaming CPUs and I went with the 9800XCD which had trouble staying in stock.
Now to read the new post...
No.4111
>>3940even more important
¥always call out a greener when it's andy
No.4112
>>4109If you have no AI plans, AMD is definitely something you can do as a cheaper and much better price point. The 9950X3D is probably something you should only buy if you already have specific plans to make use of it extensely, which most people don't. The 9800 is already really powerful at its price and that's why it's so popular.
>One of the difficulties in doing this is that everything I get has to be from a local store ...I'd like to please ask for help choosing a graphics card, motherboard and PSU, then RAM if necessary.I'm not sure how we could help since we can't go to the store with you. Maybe you can go there and take notes or something?
No.4113
>>4112Sorry, to be clear I more meant I can't order off Newegg, Amazon etc or buy second hand because the purchase is being made by my mother's company and has to be done locally. It's a chain and they have a website:
https://www.pbtech.co.nz/ Another store that sometimes has more niche products is
https://computerlounge.co.nz/. I don't know if I can buy from them in this instance however.
The 9800 sounds good. I'm sometimes a bit wary of buying a very popular part, or even something popular outside of hardware, because it often seems like it's for a common use case, but I suppose I fit such a case when it comes to PC use. And thank you for reading my post, sorry I got a bit rambly.
No.4114
>>4113Hmm... Well, I'm not too familiar with AMD cards but I'd say you want at LEAST 12gb of VRAM. Personally I don't think you should bother buying a card with less than 16gb. If you can't afford the extra $100 New Zealand bucks (or the company rather) then save until you can.
But... hmm.. I just really don't know AMD stuff. I haven't had an AMD GPU since like 2002 or something. I'm just too paranoid over driver issues with obscure emulators and stuff. It's better these days, though.
No.4196
>>2269get a radeon for the same chipset at the same price chances are it's gonna be much better
oh wait january post. how are you doing anon? enjoying your new card?
No.4197
>>4109both 5070 and 9070xt are such a weird things. they are 5 years more recent then radeon 6800xt and still losing to it in many categories.
actually, speaking of weird, lets start how fucking bizzare it is that 5060 and 5070 are base 8 gb vram when 3060 had 12 lol
No.4215
>>4196>how are you doing anon? enjoying your new card?Doing good! With AI I'm using my graphics card far, far more than I did when they were just for gaming. It used to be I'd get a decent card and then use it like 20 days of the year with specific games, but now I'm using it practically every day. I made a new SDXL NoobAI super merge to meet my demands and it took far less time since I was able to generate the images in a fraction of the time. I'm currently trying to hook up SillyTavern to SD so I can automatically generate images for scenes while the text AI is offloaded to OpenRouter. AI Video is kind of fun, but it doesn't hold the 'style' so it's not as good as I would hope. Unfortunately training video LORAs seems outside the realm even of the 5090 so I'm not sure when that would improve. People generally seem focused on porn, which is all well and good, but it quickly reverts to a generic appearance which kills the purpose of having 2D AI at all.
The only problem is AI makes it hard to find motivation to exist in real life or to create things, but I'm trying. The good news is that this card allows me to have blender open doing simple stuff unrelated to rendering, so I can AI prompt in the background while working towards other goals... slowly.
No.4216
>>4215what is the advantage of having your own ai model over using something like novelai or even chatgpt?
No.4217
>>4216Well, local models are better than online for the fact that you can customize them by adding LORAs and stuff, most local people do that. ChatGPT is of no interest to me because it's censored corporate stuff I can't use for escapism.
But when it comes to my obsessive model merging...
The model I made is a merge of like 8 other NoobAI-based merges. The advantage is selecting things I like and trying to optimize things. Some models are great at "creativity" (random mutations/errors) while others are good at poses or backgrounds or tag recognition and so on. By spending a bunch of time selectively merging different layers to try and isolate the best parts of each I try to maximize everything without sacrificing other things. For example layer OUT00 here has a huge influence on body types, so maybe someone else's model does great here, but layer OUT01 which controls the general coloring and shading is terrible, so I can take Layer OUT01 from the former and inject it into another model without the terrible colors.
No.4218
>>4217err I mean OUT00 is injected. Whatever.
No.4220
>>4217interesting. I use AI for human-level analysis. I mostly just use notebooklm, and feed it people's books, chatlogs, personal correspondence and any other source I can find, and use AI to draw this person's psychological profile, or look for inconsistencies in their narratives etc. Do you think I could benefit from having my own model?
No.4221
>>4220You're talking about text LLMs so the investment is like $10,000 to get a computer server rack setup thing that has like 800gb of fast RAM to load the entirety of the deepseek model, which has been the premiere local model for 2025 so far. Most people doing online stuff use the same model, but hosted by a company that largely charges nothing to use it. (I think they log the prompts for future training data)
The online-only models like GPT and Claude boast far larger context windows and tool use/agents so unfortunately the online stuff has a noticeable gulf in usability.
No.4222
>>4221Oh, but note that you can of course load smaller and weaker text models, but it's very much a hobby thing. People don't want to use something more prone to hallucinating when it comes to programming help and the roleplaying is far more dry since the data isn't there.
No.4246
>>4244>How can I make money from this while still being lazy...There’s some really obvious and cynical answers, but you probably don’t want to do them.
No.4250
>>4246Yeah, probably. Stuff like spamming social media or actual art sites with an AI deluge until I get enough attention?
I don't want to make the world worse, I have my basic needs met so I don't need this to survive. Probably best to keep it as a tool to help me masturbate and design characters/scenarios for games.
Sadly this low sampler setup for video doesn't actually seem to work for me and I need to do the long generation times.
It seems like you can offload some VRAM blocks to CPU (or whatever the terminology) was, but since I can barely squeak by and fit this specific resolution inside 32gb of VRAM at 12 second run time, my generation time is pretty amazing. 5 minutes for this. It would probably be about 15 hours on the 3080 if I had, if I could do it at all.
No.4436
I am now in the exact same situation and feel like my mind has already surrendered to making the worst financial decision of my life. I was originally gonna replace my 960 with a 3090 Ti, but now that I have seen just how much quicker the 5090 is for AI video alone, a feature I could actually really use, I know I wouldn't feel completely content with going for something lesser.
After doing some research, I could theoretically get one and pay in monthly installments of $100 at no interest, which coincidentally is the max. amount I usually save in a month, but I'd need to save like $50 on top of that for next year when my usual expenses increase in cost. I really need to find some loaded wallets on my weekly trip to the store..
No.4437
>>4436what do you need AI for? if you just play games, and used to be content with ancient shit like a 960
just buy radeon 6800xt or 7900xtx
Stay away from overpriced nvidia crap
No.4439
>>4438AMD didn't try competing in that area until they released R9700 a week ago. Currently it's still not available for purchase. It is inferior to 5090, but ought to be cheaper.
Rather then watching youtube, use sites like nanoreview or technicalcity to compare the specs and look at benchmarks yourself. youtubers are paid shills.
No.4440
>>4436OP here. I can say video was fun to mess with, but the time involved makes it more annoying than fun. Maybe it's just me, though. Video stuff is clearly made with realism in mind as it often struggles very much to maintain a 2D style if you tell it to do too many different things. Loops are the way to go since it's restricted to matching the end frame to the starting frame, reducing the freedom it has. If you're someone that finds 8 second loops of video far more erotic than still images then, yes, it's worth it. For me, though, it's
usually too uncanny.
ACTUALLY NEVERMIND! I forgot that a few weeks ago I spent like 20 hours messing with the ComfyUI nodes to combine two different workflows to try and make some good 2D looping. This actually looks pretty good. Well, definitely problems with her looking non-Kuon-y for a few frames, but that's why you'd generate like 20 videos and then go do something else. Alas, the randomness of AI is still here.
People keep making advances on making LORAs and similar things that allow you to generate faster, but... wait for it... it's made for real images.
Video models also use 'natural' language instead of booru tags, which I find really annoying, too. You can at least use other AI models to write the prompts or make suggestions. Video is definitely frontier territory, so keep that in mind. Also I want to say again how much I hate comfyUI.
For multitasking it's truly amazing. I can keep SD loaded and still play a game like Palworld and not feel anything, I can even have SD generating some images at the same time! Not video, though, video uses all the VRAM. If you want to do AI RP stuff with an on line text model you could also have some better stuff loaded to go along with it, like local voice model and stuff. All in all, the more you use AI the more you'll use the card.
It's kind of weird, now that graphic cards are also used for AI I use them about 5000% more often and now I actually feel like the cost is justified in a way.
No.4442
>>4438AMD GPUs are much better if you're actually interested in driving a display(s) and having support 10 years from now. You don't have to rely on AMD to build drivers against the latest kernel and it is much less of a headache to keep it going.
Nvidia GPUs are effectively a black box. You're much better off buying them used and running them as second GPU. But even then you might not be able to continue using them forever. The only Nividia GPU I own now is not really that old but they no longer build drivers against recent kernels. So I'm stuck either running an old kernel or not using it at all. I hate having to plan my entire OS around the GPU.
I refuse to support them any longer because I've gotten burned too many times. There is really no point in using AI locally either. It isn't nearly as impressive as everyone is pretending it is.
No.4446
>>4440I’m fascinated by how you’re doing this. What’s your workflow like when making these generations? Is it at all hard to make it similar to what you’re picturing in your head?
No.4449
>>4442using AI locally is great if you have actually a reason to use it
if all you use it on is generating pictures of waifus and gifs like the nvidia shill does then yeah, it's just a luxury for the rich
also another argument: price needs to be considered. if you are serious about AI, you will need to invest. ideally at least 128GB of ram and multiple gpus. Here's the question of how much cheaper the R9700 will be once they start selling is separately. Projections are up to half the price of 5090. So here's the real question: what is better, a single 5090 or two R7900s? The answer to this question will be known once benchmarks start appearing. I'd wait a few weeks with buying a 5090 until then.
No.4450
>>4449>or two R7900typo btw, meant two R9700s of course
No.4451
>>4446Holy crap someone else is interested in this amazing technology?!
I only mess with image-to-video model as I can't imagine these are any good at generating 2D stuff from nothing. It's not going to know characters and stuff, so having a starting image is a huge requirement.
My basic process is:
1. Generate a good source image in SD
2. Use AI to do a basic natural language observation of the image + make suggestions on what to animate. I'm not sure if this is doing anything, I need to test it more. For example is "She moves her left hand smoothly to pick up the glass" really any better than "hand moves with glass"? I'm guessing it isn't.
3. Generate the video with a workflow thingie that took me a few days to actually understand.
Just like with image gen there's LORAs, but they're quite a hassle. Most of them are geared for realism as I said earlier so it's not that great, but some are really really nice for NSFW stuff.
No.4454
>>4449>the nvidia shillHow much do you think nvidia is paying me to call their GPUs ridiculously expensive and to take months beating myself up over wanting to buy one?
You've entered my thread to call me a shill because I'm genuinely excited about a technology and sharing my personal experience on it to advise others. Not only that, but you're doing it purely out of ignorance.
Stable Diffusion/Image Generation?
CUDA.
Audio Generation?
CUDA.Image Captioning?
CUDA... I think? (most likely)
Video Generation? It took me a solid week to get the latest
CUDA drivers to work with sage attention and pytorch.
Local Text Generation? This one is
actually less dependant on CUDA. You can run text gen on CPU, but there's a
large speed penalty. This is how people are running deepseek locally without spending $80,000.
Everything is built on CUDA and CUDA is nvidia.
Since you're proclaiming AMD over and over, do you know the name of AMD's CUDA analog?
I'll give you a second to think about it.
It's ZLUDA. It's not their own thing. It's a hack to try and get nvidia's CUDA to run on an AMD GPU and people have been working on it for years at this point. AMD as a company is not working towards it, just random people online. AMD forfeited.
If you think I have loyalty towards a monopolistic shit company like nvidia you are sorely mistaken. I would love it if there was actual competition in AI, but that is not our present world.
>what is better, a single 5090 or two R7900s?For AI, the 5090. Unfortunately it's not even a comparison. AMD prices similarly to nvidia to such an extent that people said AMD blew it this year with its high prices, choosing to nickel and dime people instead of expanding their market share. They joined nvidia with the price gouging. Their "AI" card is being sold at a fraction of the nvidia price because it will be a fraction as useful.
No.4455
>>4454>You've entered my thread to call me a shill because>my threadno such thing, we are anonymous and this is a public place
No.4456
>>4437Aside from the obvious NSFW use, it has gigantic potential for 2D editing, think Eroge-style Motion Graphics that don't really follow a red line in terms of design. That's on top of other smaller time-saving benefits that add up over time, also related to workflow. I actually don't really play a lot of games myself.
>>4440Thank you for your input, I was planning on asking!
What are your observations regarding the progression / optimization of local generation, does the community make an effort to try and make models less resource-intensive over time? Or is it similar to unoptimized AAA games with huge raw assets no one cares to compress, instead expecting the consumer to upgrade early? Because that's my main worry here. As someone who doesn't use camera footage, I know I will be set for non-AI editing for a decade+, but I haven't really been able to get a feeling regarding AI. For example, does image generation nowadays use less VRAM at more speed and equal / good results on, say, 20-series cards?
Also, what's up with the supposed lack of 2D-focused models? I think to remember that the threads on 4/h/ (might've been /trash/ instead, it's been a while) had something relatively exciting to report about every few days. To be fair, that was when the AI craze was still new and it was only related to images.
No.4457
>>4456>does the community make an effort to try and make models less resource-intensive over time?Yes and no. Models themselves can't be made more efficient, it's basically a settled technology once it gets to the hobbyist's hands. It's the corporations and research teams that could make more efficient models, and then some rich guy could rent a data center to create a 2D model based on that tech. But he's not making a new technology, just training one based on another.
There have been a few efficiency improvements over the years in regards to VRAM usage and stuff, but I don't think much of it is
new new, it's a few years old now. There's probably some gains for image gen dependant on ComfyUI, but I HATE ComfyUI and do everything I can outside of it. I know for video there's a VRAM offloader thing in ComfyUI that increases generation time but allows you to offload some stuff to RAM so you can generate the thing at all. I use it even though I have 32gb of VRAM. Video is THAT demanding once you get to certain sizes or lengths.
Anyway, those VRAM gains for image gen are offset by the newer models being made to generate larger images, which requires more VRAM. People also want to use the newer bells and whistles like upscaling and adetailer and controlnet and other stuff which each have their own VRAM cost. But you can definitely enjoy image gen with SDXL with 8GB or possibly even less if you heavily restrict active programs and limit yourself to maybe 1-2 images generated at once. I still have hardware acceleration disabled in all my browsers to free up a few hundred megs of VRAM.
There have been more improvements to LORAs/models that can generate things in fewer steps, which means less generation time which is most noticeable on people with weaker cards. However it generally comes at a cost somewhere, usually quality, and I never liked the ones I tried. (and the ones I didn't try require manual installation by altering code and that's above my brain). For video models, these faster generation things come at a cost not just of quality but they also tend to heavily restrict the animation in various ways, so each one I've tried left me disappointed. Except "lightx2" LORA, that one seems good and it's in all of my recent video things.
>Also, what's up with the supposed lack of 2D-focused models?There's no money in it I would assume, or at least comparably. For video it's probably just that RL video data is far easier to come by when it comes to data scraping. And if somehow your dataset lacks something, grab a smartphone and record it yourself.
No.4458
>>4457Man, how disappointing about the excessive demands for video.. even if your 5 minute result from above does look very promising. What do you assume a 10s video would take to generate if you go for the maximum resolution that still fits into your VRAM with no offloading?
I stopped experimenting around the time people talked about LORAs, I just used the so-called "NAI leak" back then and decided that waiting 5 minutes for a single super low resolution image with horrible quality isn't worth the hassle. I've always wanted to get into training models with my favorite artists, but that's completely out of the question with my current card. I shall see if it's really gonna be the 5090, might just be the best "budget pick" on Nvidia when it comes to VRAM, though I haven't checked past workstation cards yet.
Either way, thank you very much for all the info! I shall make great use of it in the future.
No.4459
>>4458>Man, how disappointing about the excessive demands for video..That's just for the extremes. 720x960 is quite huge for a video size and then the length greatly increases the VRAM usage. I can do 81 frames at 720x960 at like 22ish GB of VRAM.
My 'block offset' is 9 blocks (moved to RAM) when I do 161 frames, which is certainly on the extreme end of things. Videos tend to get errors with the colors at this length so I don't often try it. It might be a workflow thing, though.
>What do you assume a 10s video would take to generateIt goes by frames and not seconds, with adjustable framerate. 12 or so FPS seems to be about the slowest I can go without consciously noticing it's slow. 81 frames seems to be sweet spot, possibly what the model was trained at, so it would be a bit under 7 seconds.
No.4460
>>4458>What do you assume a 10s video would take to generate if you go for the maximum resolution that still fits into your VRAM with no offloading?Oh and I'll time and test this later, but can't right now. It should be like 5-10 minutes when I push it to the limit of frames.
No.4477
>>4460Okay, I did some testing. I'm using 12 frames a second.
720x960 resolution and 81 frames: 25gb of VRAM used and it completed in
171 seconds. I forgot that NOT offload blocks to RAM, but I assume the gains wouldn't be that major. Maybe a dozen seconds?
Don't know why it's behaving so weird with the colors. Such is video AI.
No.4478
>>4477161 frames. 9 blocks offloaded to VRAM. 29gb used.
317 seconds. Maybe it's the pure white BG that's making it so weird.
No.4481
>>4459>>4460>>4477>>4478Thanks so much again! It's nice to see how the first example seems to try being more creative with the animation. Most of the videos I have seen have this "animated PNG" feeling to them, similar to Gacha games. Have you read anything about RAM speed making a significant difference when offloading?
Also, would you mind sharing your sources when it comes to AI news that revolve around models designed for us 2D-focused people? Or really anything that would still benefit that genre.
I sometimes stumble upon random YT videos that showcase a large variety of things that seem incredibly interesting (and increase my 5090 itch), but it's just so much at once.
Last thing: We need more Koupen!
No.4486
>>4477>>4478Bleh, I now notice mistypes. I mean it's loaded
from VRAM and not
to it. The "offload blocks" thing moves it to RAM or something to reduce VRAM usage.
>>4481>Also, would you mind sharing your sources when it comes to AI news that revolve around models designed for us 2D-focused people?I just check civit.ai now and then to look at models directly. AI news, particularly youtube, is full of scammers and opportunists that promise you the moon so you'll click their thumbnail and are the same scum that did (and sometimes still do) the crypto hype stuff. I'll also rarely scan AI generals since it's easy to find informative posts as the guy will usually type more than 8 words, which makes it stand out from 95% of posts inside a general thread.
No.4762
>>3893>I feel immense guilt and shame over itAs you should! For ~$6K you can now get a DGX-1 system with 256GB of pooled VRAM. If you had saved your money
instead of buying a 4090 and then a 5090 we could have an actual Kissu AI server. Tsk tsk tsk.
https://www.ebay.com/itm/167512048877
No.4763
>>4762I figure this is how AI stuff will go if you want to be serious about it. Buying a GPU and building a desktop around it isn't the best idea. If you get an Nvidia GPU now you'll have to deal with them dropping driver support in 5-10 years. This really sucks on Linux because you end up stuck on an old kernel or running the terrible FOSS drivers.
You're better off buying a GPU that provides full driver source code for your display devices and building a server for working with AI stuff. It's designed for that anyway. You send data off to be processed and it comes back.
Either that or wait until they start providing dedicated AI chips. I figure that is coming down the pike in the next few years. Built right into the motherboard (or at least a socket for it). I see us going to unified RAM soon instead of having the current RAM/VRAM split. Maybe this will improve the situation with buying a card to run displays and real time graphics. Right now things are so bad I've put off buying a new GPU for almost a decade. I can't justify spending what they're asking when my old one still does what I need.
I've moved most of the things I used to do on my desktop off to servers. All my files are stored there and it's running several things like a music+video daemon that can serve the data with metadata to whatever other device I want to use. Over the LAN and the internet. It has been much cheaper in the long run and having access to everything when I'm away from my main desktop has been very nice. It reduces how much I spend on other things like laptop and small handheld devices.
Prices are going up on the stuff you need for that to. But things like enterprise switches+routers+serve
rs with server grade CPUs+RAM+HDDs are very cheap if you shop around. They dump so much of that on to the used market every year. The only thing that's really expensive still is the stuff you need to run a fiber network inside of your home instead of on CAT5/6 cabling. Which can push 10Gbps anyway and will probably get another speed increase with time.
No.4771
>>4762It does seem cool, possibly, but I'm beating myself up after scraping together $2800 so spending $6000 is simply impossible.
I'm really not sure how or
if that thing you linked could be used for my AI purposes. From what I've seen having two GPUs instead of one complicates things, it's not just a plug and play kind of deal. You load THIS into GPU A, and THAT into GPU B, but you need to merge the result HERE and so on. I'm not some machine learning guy, I just run what other people have made.
The crux of the matter, though, is that this stuff is only as good as its adopters.
Right now there's been good progress in local video stuff because more people are getting 24gb and higher VRAM, which allows the hobbyists to create and share things. Without a community of people with the hardware it's not that useful.
As for text LLM stuff, 256GB is unfortunately still too low to load the best local models. There's a massive gulf and mid-range stuff largely doesn't exist.
No.4919
>>4771>still too low to load the best local models256GB should be enough to run GLM-4.5-Air at FP16 (225GB), or GLM-4.5 at around INT4 (~190GB; Full-size is 755GB at FP16). Air has rough parity with GPT4o-mini, and GLM-4.5 rough parity with GPT4o. That's purely for knowledge-level tasks, though. No idea whether they're any good for RP or whatnot.
No.4992
>>4919Well, that's good to hear. That's still extremely out of my price range, though. Maybe we can start a Kissu Fund where everyone gives me thousands of dollars. What's the context window with that kind of setup?
That sounds really, really nice. Well, I guess it depends on the (E)RP potential, though. Ultimately people want the best of the best for things that matter, like programming, but ERP can be 'eh, good enough" if you're in the mood.
No.4994
>>4992>What's the context window with that kind of setup?Probably in the range of 5-15K tokens... So... Usable with aggressive summarization, but maybe not so great for long, detail-oriented back and forths. Something something.... you can quantize the LLM, but not the context window (KV cache)?
Anyways, you can actually use GLM-4.5-Air and GLM-4.5 on
https://chat.z.ai/. It's a bit faster than Deepseek, but nowhere near as fast as ChatGPT.