[ home / bans / all ] [ amv / jp ] [ maho ] [ f / ec ] [ qa / b / poll ] [ tv / bann ] [ toggle-new ]

/maho/ - Magical Circuitboards

Advanced technology is indistinguishable from magic

New Reply

Options
Comment
File
Whitelist Token
Spoiler
Password (For file deletion.)
Markup tags exist for bold, itallics, header, spoiler etc. as listed in " [options] > View Formatting "



[Return] [Bottom] [Catalog]

File:Medalist.S01E01.A.Genius.o….jpg (327.23 KB,1920x1080)

 No.2269[Last50 Posts]

I'm feeling really, really tempted to get a 5090 even though it's a massive scam and it's months worth of savings. It's not the rational thing to do, but I do AI stuff a lot and it brings me joy. (no I don't condone AI shitting up the internet and art)
There's also 3D modeling I want to take more seriously after Palworld reignited my passion for building stuff. Obviously you don't actually need a top of the line card to do this stuff, but it does allow more geometry to be active and speeds up rendering massively. More VRAM means you can have more processes open so jumping between programs is smoother.
AAA gaming sucks apart from Capcom so that doesn't really enter the equation at all. I guess ironically retro pixel filters are known to be VERY demanding if I decide to do that. Might be more demanding on CPU, though, I can't remember.
I'm in that CG tracker that went private 5 years ago so software and assets are no issue, but man this is still such a huge amount of money.
What to do...................................................................................................................................................................

 No.2270

.flip it

 No.2271

File:Medalist.S01E02.Prelim.Bad….jpg (524.73 KB,1920x1080)

>>2270
Do NOT say that about Inori!

 No.2272

>>2271
... no I meant flip a coin for it, you can either follow results or see vid related.

 No.2273

I might buy a used 4090 since everyone's dumping them on the secondary market locally and the 5090 is shit.
I'll only give the envious green AI snake oil company money indirectly at this point.

 No.2274

why are you buying an nvidia card. nvidia is an AI for buisness company

 No.2275

Couldn't you get like two 3090s or two 4090s (or some other combo) used for the price of a new 5090? that would probably be better for AI than a single 5090. Or maybe a quadro and the likes. I wouldn't trust the 50 series cards just on principle that they lied about the performance metrics of the 5070, saying it's better than the 4090 at half the price.

 No.2276

File:1462665516599.jpg (8.28 KB,178x178)

Is there any way to do 3D modeling from the kissu pc because I'm already for sure getting a 5090 for that.

 No.2277

>>2276
There's a kissu pc?

 No.2278

>>2277
Yeah it's what the Palworld server is hosted on

 No.2279

I believe there are ways that it's used in LAN. The latency would probably be pretty bad by network

 No.2280

Really it's about rendering rather than as a studio. Rendering for 3d is only ever done by distributed systems

 No.2281

Pathetic

 No.2282

File:[SubsPlease] Salaryman ga ….jpg (269.89 KB,1920x1080)

>>2275
Two GPUs unfortunately have a hidden cost to them. I'd need a new motherboard and if I need a new motherboard then I need a new CPU. I might already need a new PSU if I'm getting a 5090, so the costs are absurd. Bleh!

>>2276
Not really, or at least I doubt it. That kind of setup is for rendering farms, once you have a complete scene and you send the data over to be processed, like a Pixar movie. The performance I'd want is for rotating the camera around without stuttering in a complex scene while having other 3D programs open or even a game window. You know, "live" stuff. Multi monitor setups are less impressive when you can't make use of them for everything.

 No.2285

i don't even know if multi-gpu is a thing that works. when I tried to do it for video games it always had issues and the data had to be mirrored across both cards.

 No.2286

>>2285
Don't know about gaming, but that's how they do it at the big AI training servers with 100*s of GBs of VRAM.

 No.2309

File:NVIDIA GeForce RTX 5090 Fo….png (8.91 MB,3840x2160)

The 5090 doesn't seem dramatically better compared to the 4090, like the 4090 was compared to the 3090/Ti. It's mostly just a generational improvement. Real world performance seems to be about 30% higher. That performance increase seems basically identical to the increase in TDP; 450W on the 4090 compared to 575W on the 5090, or ~28% higher. The efficiency in terms of FPS per watt more less bears this out, with the 5090 being around the efficiency of the 4090, or slightly below.

In Gamers Nexus' testing (https://www.youtube.com/watch?v=VWSlOC_jiLQ), for reference, the 5090 had an FPS/W of 0.34, and the 4090 at 0.35. The 5090 drew 538.4W, and the 4090 drew 391.7W.

Not exactly very impressive. It basically looks like the 50-series is going to be a refresh generation.

 No.2310

>>2309
What about for AI, like most people have been touting it for?

 No.2311

File:Screenshot 2025-01-23 at 2….png (42.99 KB,577x761)

>>2310
Hard to say. AI compute could be anywhere from 1.3x to 2.5x greater than the 4090. You'll likely have to wait until the release date on January 30th when regular people get their hands on one.

If you believe Nvidia's marketing, it should be 2.5x the speed of the 4090, going by their reported Tensor core TOPS. But... Well... They also said the 5090 would be twice the performance of the 4090, which it very obviously isn't (>>2309)...

 No.2313

File:TASOKARE.HOTEL.S01E03.The.….jpg (260.86 KB,1920x1080)

>>2311
Yeah, I'm on at 3080 with 12gb of VRAM at the moment. It's a massive upgrade for AI just based on the VRAM alone. For a lot of the AI stuff you need a certain threshold of VRAM just to run it, like the video stuff I've mentioned before would take me 30 minutes to do what a 4090 does in a minute, and that's with making a lot of sacrifices. You can either store it in speedy VRAM or you can't.
>January 30th
Nice, I still have some time to think it over.

 No.2315

File:Screenshot 2025-01-24 2134….png (425.28 KB,1920x1080)

>>2310
>>2313
Aha! One of my favorite tech YouTube channels I watch did an AI benchmark. It looks to be between a 1.3x and 1.45x improvement over the 4090, similar to the gaming uplift. The review did mention something interesting, however, which was that if the model in question leveraged INT8, then you would see a 2.5x improvement over the 4090 like you would expect just by looking at the AI TOPS, but they didn't show any benchmarks for this.

https://youtu.be/nryZwnVYpns?t=1615

 No.2316

File:Medalist.S01E02.Prelim.Bad….jpg (238.59 KB,1920x1080)

>>2315
Nice, thanks. The two yotubers I look at once in a while (gaming nexus and jayz2cents) just looked at the gaming side of things. I'm generally more interested in image/video AI more than text since text has such absurd VRAM requirements for the better models. I'm sure the efficiency there is increasing over time, or at least I hope.
I'm leaning towards getting the card, but I just learned that it's a PCIE 5 thing instead of the 4 my motherboard has. 5 is backwards compatible, but I won't get the best possible performance. I guess I could buy the card and get a new motherboard in summer or so.
No reports of its performance in 3D programs, though. I'm sure it's good, but I'd like to hear more about it.

 No.2356

File:[SubsPlease] Fugukan - 03 ….jpg (319.52 KB,1920x1080)

Yeah, I'm going to do it. I've got one life to live and blah blah blah. If this goes wrong then at least my corpse will have a smile on it.

But it's going to be more expensive than I thought, but thankfully I can do it in two increments. I need a new power supply for this and that has to happen now. However, my motherboard is PCIe 4 when the 5x series is PCIe 5. They are thankfully backwards compatible, but performance suffers. I guess I can see how bad it is, but if it's BAD then that means I need a new motherboard which means new CPU and new RAM. These are about 4-5 years old so I guess it's not unreasonable for them to be upgraded, but I don't feel like I need them so it's a bummer.

 No.2357

oh wait, now I see that I said something similar 2 days ago
being sick is fun

 No.2406

File:[SubsPlease] Jibaku Shoune….jpg (456.13 KB,1920x1080)

It seems like Best Buy is only selling them online on the site itself which means I'm forced to compete with bots and everyone else on the planet. Yeah, this isn't looking good. I'm just going to assume I won't be able to buy it, but I'll check every hour.
Maybe this is for the best.

 No.2407

File:[SubsPlease] Jibaku Shoune….jpg (353.48 KB,1920x1080)

Yeah. I hit the 'add to cart' button easily within the first minute it was up, but the scalper bots got them all first. Well, I guess the decision was made for me then.

 No.2412

>>2273
Update: I think a lot of people had the same reaction to Blackwell being shit, all the local 4090's disappeared just a little after the official reviews broke.

 No.2478

You can't even buy graphics cards nowadays without business connections or bots

 No.2479

>>2478
Yeah, he's not the first to call it a paper launch. I'm not even sure what my plans are now.

 No.2480

>>2479
if you've got 1000 dollars to burn buy a vr set and get a better idea of what you can do with 3D models+inspiration.

 No.2481

>>2480
VR is something I looked into for easy money once, but it's not something I'd want to deal with. The first thing is that it's Unity and I don't want to learn an engine whose primary focus is phone games. The second is that I'm not generally good at talking to new people and making ERP avatars for people is very personal.
Has anyone said anything about more 5090s coming in at any point? Quarter 2 or summer or anything?

 No.2483

>>2481
also Unity is the engine with that guy that wanted to monetize it further and everyone rallied against online. If possible people should avoid Unity since he's just biding his time.

 No.2486

>>2483
sounds like a minor issue

 No.2487

>>2486
It really isn't a minor issue that a game engine seeks percentages of all your profits and wants to attempt to retroactively change things. It's a massive liability.

 No.2488

>>2487
it's minor because there's UE, or directly interacting with the XR APIs with software.

https://vroid.com/en/studio is where models are uploaded and "supports Web (API), Unity, and Unreal Engine."

 No.2494

File:[Piyoko] Himitsu no AiPri ….jpg (298.8 KB,1920x1080)

I think I'm going to buy the other parts now.
I picked a great time to do this apparently since the 9800XD is a very popular CPU that came out recently and... it's out of stock everywhere.
Man, I just want to upgrade my computer to something that will last a good 5 years.

 No.2498

File:[SubsPlease] Fugukan - 04 ….jpg (228.73 KB,1920x1080)

I think I picked the worst time in years to upgrade my computer. This is just dreadful. GPUs are completely out of stock. CPUs are completely out of stock.
I'm hesitant to look at motherboards or memory right now because my assumption is there's something going on with them, too, and I can't really get those without a CPU.

 No.2499

>>2498
I mean, pretty much everyone and their mother was waiting for the 50 series to come out before upgrading...

 No.2500

File:[enm] Medalist - S01E05 [W….jpg (286.9 KB,1920x1080)

>>2499
New computer parts come out every year. This one wasn't any different, apart from nvidia bringing 500 GPUs total to market.

 No.2501

File:[MoyaiSubs] Mewkledreamy -….jpg (312.12 KB,1920x1080)

Actually, does anyone have recommendations on how I go about finding a good motherboard and RAM? It's a lot easier to do CPUs and GPUs, but this stuff has a lot more options from many different brands. I've never really felt like I've done enough research when buying them.

 No.2502

There's a table for comparing MHZ and cas latency. There was some other things about the number of chips on them or something.idk.
For boards... I think transfer speed is pretty fixed and it's down to stuff like ddr and pcie versions. And picking the best of a few brands..

 No.2503

File:[enm] Medalist - S01E05 [W….jpg (352.93 KB,1920x1080)

CPU SECURED! (I hope)
If somehow anyone else has been refreshing sites hoping to get the 9800X3D, it's on amazon! And it's directly available from amazon instead of mysterious chinese shops created days earlier.
https://www.amazon.com/dp/B0DKFMSMYK

 No.2512

File:1739649080125935.png (83.34 KB,1955x524)

Saw this on /v/, so a /v/ post about a reddit post about a twitter post.
I think I can trust it. "Bap" is in the name

 No.2529

File:cpu.jpg (868.74 KB,1728x1300)

I'm going to turn this into 'assembling a new computer' blog.
The CPU arrived.

 No.2532

>>2512
https://www.tomshardware.com/pc-components/gpus/rtx-5090-supplies-to-be-stupidly-high-next-month-as-gb200-wafers-get-repurposed-asserts-leaker
Seems like it's taken seriously enough to appear on some hardware news sites. So, 5090 next month before economic uncertainty.

 No.2543

File:aio.jpg (444.3 KB,1415x800)

CPU All-in-One water cooling thing is here. It's highly rated and it's actually cheap for water. $50. So cheap that people suspect it's a loss leader thing.
It seems like AIO's are more important for these giant modern GPUs that block so much air flow, you can't rely on direct cooling via a heatsink that's so close to the slot.

 No.2544

File:ram.jpg (978.07 KB,1512x1432)

RAM is also here.
I'm going to bump to let people know this info here about 5090s likely being in high supply in a month: >>2532

 No.2545

>>2532
Mmmmmm, seeing some conflicting info about the melting wires here. Is it something to be worried about on any setup, or are there easy ways to avoid it?

 No.2546

File:Medalist.S01E07.Level.One.….jpg (449.53 KB,1920x1080)

>>2545
There's also a thing about some cards not having sufficient ROPs and I don't know what a ROP is exactly but it means weaker performance:
https://videocardz.com/newz/nvidia-confirms-0-5-of-rtx-5090d-and-rtx-5070-ti-were-shipped-with-fewer-rops
It's something you could check yourself and they'd replace your card for it, though.
If the news is true and they're pivoting to making lots of 5090s then I could see them looking more closely at the manufacturing process. The wires thing seems pretty rare, just like the 4090s, but who knows.

 No.2548

>>2544
this meower just rammed me

 No.2576

File:mb.jpg (838.07 KB,1867x1249)

motherboard

 No.2577

File:case.jpg (1.06 MB,1948x1404)

case
cat was sleeping

 No.2578

File:0.jpg (1.37 MB,1925x1323)

The case is quite nice, and it better be at the price. ($180, but I justified it because I buy a new one once a decade).
Lots of nice slots and it has a little plastic case for screws. It has a fan controller chipset thingie which is very helpful, I used to have one I bought separately and fit in a CD tray. Cases are quite roomy now without optical drives, huh.

 No.2579

File:1.jpg (1.28 MB,1928x1336)

Replaced CPU AIO's fans since they're said to be loud. Got a 5 pack of thermaltake ones for $30.

 No.2580

File:2.jpg (441.86 KB,1008x756)

Ram in motherboard, motherboard in case.
The motherboard is known as one of the cheapest X870 boards (ASRock X870 Pro RS). It performs well, but and they really skimped out on everything other than the motherboard. You don't even get a booklet. You get ONE SATA cable which made me laugh. I have to look online as to which slots are what, like the M2 slots. Well, I guess it's a good way to cut costs.

 No.2581

File:bonus.jpg (2.24 MB,3024x1875)

Bonus shot of one of my storage containers since I'm looking for an audio optical cable thingie. The new motherboard doesn't have connectors for the 3-way split audio cable. I hope I don't have to order a new one, but I think I might. I wish I held on to all my physical boxes for PC games...

 No.2582

File:asdf.jpg (1.33 MB,2419x1651)

Searched a lot and found no cable, so I ordered one. It's just $8, but it will take a day or two to get here.
In the meantime, time for a drip down memory lane!
My mom actually bought me the FLCL and EoE things, I've yet to this day pay for anime. Thanks, mom, maybe DVDs will be worth something one day.

 No.2583

File:asdfg.jpg (1.46 MB,2327x1814)

Someone build me a time machine because I want to go back.

 No.2584

File:asdfgh.jpg (1.43 MB,2393x1697)

OK that's enough off-topic blogging for now. Time to go back to assembling the computer, but I'll have to move parts from my existing one to it so I'll be offline for a bit.
Video game booklets are something I want to start collecting sometime, though, they're just so cool. I wonder if anyone actually sells them, though, or if people assume they have no value.

 No.2598

File:blehh.jpg (998.91 KB,1901x1335)

moved the re-usable stuff from old computer to new computer...

 No.2599

File:blehh3.jpg (757.59 KB,1397x1869)

but my yahoo mail accounts (yes I still use it for old account stuff) told me to confirm my identity and I had a circular thing going on where to confirm account A I had to use account B and to confirm account B I had to use A, so I had to partly re-assemble the old one to fix that and then re-assembled again.
That wasn't very fun, but I enjoyed the stuff before that.

 No.2600

File:blehh2.jpg (644.77 KB,1789x1397)

Glad I still kept my old power supply so I didn't need to re-wire all that stuff.
(this neon glow is overboard and I'll remove that in the bios tomorrow)
Also it's kind of bothering me that the bottom fans don't match, but half the computer will be hidden behind a monitor anyway.
NEW COMPUTER YAY!

 No.2601

>>2600
cool PC!

 No.2602

>>2600
oh and I did fix the Thermaltake logo thing, the top part pops off.
Ran Prime95 for an hour and CPU and memory didn't crash at all.

 No.2603

>>2543
>water
spooky!

 No.2604

File:1593467016008.jpg (353.3 KB,1280x720)

>>2603
It's important to keep your computer well hydrated!

 No.2606

File:61dXffEuFuL.jpg (76.71 KB,1500x1500)

The CPU cooler is using an "All in one" system as mentioned here >>2543
It's water cooling, but it's a closed system for one part. It's sealed up in the factory so I can't do anything with it. The benefit is there's really no chance of leaking, but the con is that it will very slowly evaporate over time and I can't replace the water. That should take years, though, and if it doesn't then at least it was the same price as an air cooler. (This one is highly rated)
Part of the reason I did it is that the 5090 will be massive and expel hot air towards CPU, so having a radiator at the top of the case for cooling will greatly benefit the CPU.

 No.2624

neat bit of "insurance"

 No.2636

>>2624
That's pretty cool. It uses a vacuum or something? Back when I was doing my hobbyist watercooling stuff back in the 00s you would spend a few days running the loop with a separate PSU to check for leaks, but leaking almost never seemed dangerous to the computer since gravity naturally dropped the stuff away from parts.
The failure points on this AIO >>2600 is where the tubes connect to the radiator in the top right (out of picture) and from the CPU block (the upside-down 'thermaltake' square) and in both cases the water would be falling straight down to the bottom of the case. GPU blocks similarly would be on the bottom of the card.
Water and electronics are definitely a risky combination, but it's not as bad as people think.

 No.2638

>>2636
>That's pretty cool. It uses a vacuum or something?
If I remember right, it uses two separate pumps (one on the inlet and one on the outlet) and when the resistance falls due to a leak occurring, one of the pumps reverses flow so that the liquid is pulled into the reservoir.

 No.2645

>>2638
I was thinking it probably doesn't help well with punctures until a certain amount of liquid has spilled

 No.2646

>>2645
Yeah, I think it mainly makes sense as a prevention against server techs killing your $150,000 AI server (and every server below it) if they accidentally dislodge a tube off a barb fitting or something.

 No.2649

File:m2.jpg (462.68 KB,1797x1257)

For some reason the new motherboard isn't reading my oldest M2 drive which is what had my profiles on it for browsers and stuff.
Bought a $20 external M2 SSD reader thingie and... thankfully the drive is working. It makes me wonder what is wrong with the motherboard, though, maybe the M2 is too old? I moved it to different slots and even tried removing another one to see if somehow it was a bandwidth thing or something.
Kind of annoying, but at least it's a small drive and I could put some less accessible stuff on it like older AI models or something.

 No.2666

File:4e7ab9017cee87d2748f6de73e….jpg (337.59 KB,1031x728)

absolutely incredible cats in this thread
a confluence of cats and computers
the cat blog sends you its regards

 No.2818

File:Even.Given.the.Worthless.A….jpg (191.5 KB,1920x1080)

I've been monitoring 5090s and it seems like most of them getting listed online from the third party resellers (MSI, PNY and other companies) have at least a 25% markup, selling for at least $2500. Some of them, with identical hardware mind you, is selling for over $3000.
I just saw some get listed on Wal-Mart for $3900 and people actually rushed to buy them up. This is a card with a retail value of $1999 which is itself already ludicrously absurd.
I have to laugh because life is too ridiculous.

 No.2819

File:firefox_QrhLTNmL7M.png (518.27 KB,1480x946)


 No.2820

>>2819
>$5500
there are places where you could buy an apartment for the price of a few of these

 No.2821

File:>o>.jpg (127.52 KB,381x374)

>>2819
>5500
That's a little over $1/RTX!

 No.2822

>>2820
That price is actually pretty close what i bought my previous apartment for. Then again, that was like a decade ago and it's probably several times more expensive now, so i guess your point still stands.
Imagine putting "new graphics card" as the reason for a mortgage application...

 No.2828

File:waterfox_3j0NJ5XF4B.png (87.33 KB,1100x259)

These restocks that aren't at 7AM are really my only chance and I missed it. Oh well.

 No.3893

File:money gone.jpg (661.66 KB,1818x1488)

I feel immense guilt and shame over it, but I have it installed now alongside a PSU I "reserved" 3 months ago when it was out of stock. I figure spending the extra money is worth having the card this year instead of the next as I work towards things.
Sadly, my wonderful and beautiful computer parts model is no longer of this world. It puts things in perspective that time is limited.

 No.3894

>>3893
And yes, it has Ice Storm AND Fire Storm. I don't know how they don't cancel each other out, but that's the power of modern GPUs.

 No.3895

File:R-1747624704974.png (3.21 MB,1920x1080)

>>3894
Modern GPUs have tons of MP (almost 75)

 No.3896

>>3893
>Sadly, my wonderful and beautiful computer parts model is no longer of this world. It puts things in perspective that time is limited.
;_;

 No.3926

File:Over 15 Kuons a Minute.jpg (406.88 KB,1089x1199)

I was able to prompt 50 960x1088 images at once. Generation time was a little under 3 minutes with VRAM spiking at only 22GB when the part at the end with the VAE happened. I'd guess I could probably prompt 80 or possibly even 100 at once! But, it's already a pain to navigate the generated image thing afterwards so I'd never do that. 50 is already annoying enough.
I guess this is technically something possibly with a 24gb card, although that's a very slim margin.
Soon I'll move onto training a LORA for these new models! While Kuon looks great in these, it's very difficult to separate her as a person from her clothes. The (Utawarerumono) necessary in her prompt also receives some heavy influence from Eruruu.
Training should require maybe 10% as much time as on my 3080's 12GB.

 No.3927

>>3926
That's a lot of Kuon, almost too much. Your Non-Fungible-Kuon stock is plummeting in value!

 No.3928

File:05421-masterpiece, best qu….png (1.22 MB,960x1088)

>>3927
The 'wildcard' system I have setup to queue up random stuff in the prompt works very well! I need to add a lot more stuff, but the results can be really fun. This is such a wonderful use of technology. Imagine endless randomized Kuon outfits in endlessly randomized accessories and randomized this and randomized expressions and randomized that and wow!

 No.3929

>>3928
but it really does bother me how it doesn't get her tail and ears correct all of the time!

 No.3930

>>3928
>Imagine endless randomized Kuon outfits in endlessly randomized accessories and randomized this and randomized expressions and randomized that and wow!
you'll love the Utawarerumono gacha

 No.3931

File:E38195E381A3E38195E381A8E8….png (704.59 KB,1200x1920)

>>3930
I am somehow familiar with it since I 'follow' the art uploads on an exhentai gallery. I'm not a fan of how it spits all over the story with time travel and other dimensions and the other gacha justifications to commodify the emotional connection to beloved characters.
I ignore everything about it apart from the nice art.

 No.3932

>>3930
>he fell for the gacha blunder
Never spend money on gacha when the art is free and you can make your own combat in your head!

 No.3934

File:fut.png (106.95 KB,240x240)

nobody needs anything more than a 1070

 No.3935

File:96785844_p7.jpg (627.67 KB,1480x2152)

>>3932
Who are you quoting?

 No.3937

>>2269
>I'm in that CG tracker that went private 5 years ago so software and assets are no issue
cgpeers? Can you see if something I've been looking for is there?

 No.3938

I'm currently in the process of upgrading to the best hardware for editing + AI that my budget allows, which would be a 3090 Ti. I'm not too worried about not getting the 5090, but what hurts me is that I'm not able to afford a 4090, because the leap from the 30 series to it seems to be very significant.
I usually only do a full upgrade every decade or so, so I'm checking for motherboards with two PCIe 5.0 slots just in case I might be able to afford one or two capable cards for cheap in the future. I've found some 690 / 790 MBs that I might be able to get for around 150 bucks, if I get lucky.
For the CPU, I'm currently looking at the i9-12900KS, with the possibility to swap it out for a 14th gen in the future. It seems that the 14900KS will be the last for this socket, so I limited my RAM options to its 192 GB and DDR5 5600 maximum, even if the 12th gen I'm aiming for right now can't make use of it.

I'm still somewhat new to researching hardware and only ever got random pre-builds in the past, so I'm open to any criticism if something I said is actually nonsensical and I'm unaware of it. I tried my very hardest to understand how the hardware might interact with each other and to take it into account for possible seperate upgrades a few years down the line.

 No.3939

i boughted a 4090

 No.3940

>>3935
One of the two classic blunders that one must know, the second, lesser known is this:
>Never go against a green texter when funposting is on the line.

 No.4109

File:Speccy64_aUCYD5A40l.png (28.59 KB,605x457)

I'm not very good with hardware compared to most enthusiasts and was hoping /maho/ could please give me some help with something of a new build. I've been offered a free PC upgrade through my mother's company but I'm not confident in what I'm doing. From what I can read at the moment, AMD has overtaken Nvidia for practical affordability and price:performance ratio, but doesn't offer very high end GPUs. I've also been recommended an X3D CPU by an acquaintance, but I don't believe Intel offers an equivilent yet.

I've been using Intel+Nvidia all my life just because it's what I had help with from my older cousin growing up. My main use case is games, as well as some Adobe suite work mostly Lightroom and Photoshop and a small amount of video editing. I don't do anything in the vein of AI, and don't really play many new release games; my current hardware is able to run most of what I play without issue with the most demanding thing being Baldur's Gate 3, heavily modded Skyrim, or Team Fortress 2. Although TF2 is an older game it struggles on my current hardware and I believe simply upgrading the CPU will do wonders for it. I intend to play TES6 when it eventually releases.

https://nz.pcpartpicker.com/list/vgZ9Lc

I am already using the case, CPU fan and RAM in the above link. From what I can understand the 9800X3D and 9950X3D are functionally the same for my use case with the latter just being better for workstation use. According to this website and this review

https://www.videocardbenchmark.net/video_lookup.php?gpu=Radeon+RX+9070+XT&id=5956
https://gamersnexus.net/gpus/amd-radeon-rx-9070-xt-gpu-review-benchmarks-vs-5070-ti-5070-7900-xt-sapphire-pulse

The 5070 Ti is slightly better than the 9070 XT at a slightly higher price point, but is overpriced compared to RRP.
One of the difficulties in doing this is that everything I get has to be from a local store called PBTech due to use of a business account to recieve a discount, so something like the 7900 XT is not an option. I'd like to please ask for help choosing a graphics card, motherboard and PSU, then RAM if necessary.

 No.4110

>>3937
>cgpeers? Can you see if something I've been looking for is there?
Sorry I didn't see that post until now because someone bumped the thread. kissu has been too fast and active!
Yes I can, what are you looking for, person two weeks in the past?

>>3938
VRAM is the biggest thing for AI. Sure, the 5090 and 4090 will have some speed over the 3090 but having the 24gb of VRAM on the 3090 is what allows you to do many things at all. A 3090 is like having a $400,000 Ferrari instead of a $500,000 Lamborghini. It's not an exaggeration that the change from a model being offloaded some to RAM from VRAM is like 30-50x slower, if it's possible at all.

>so I'm checking for motherboards with two PCIe 5.0 slots

Be aware that the newer cards are huge. I can't physically fit my old card in the motherboard on the second slot. I wonder if someone will make an aftermarket cooler that's thin, but I bet it will be expensive as all of these things are.

>For the CPU, I'm currently looking at the i9-12900KS,
When I did research for CPU stuff AMD was the kind of mixed work/gaming CPUs and I went with the 9800XCD which had trouble staying in stock.
Now to read the new post...

 No.4111

>>3940
even more important
¥always call out a greener when it's andy

 No.4112

>>4109
If you have no AI plans, AMD is definitely something you can do as a cheaper and much better price point. The 9950X3D is probably something you should only buy if you already have specific plans to make use of it extensely, which most people don't. The 9800 is already really powerful at its price and that's why it's so popular.

>One of the difficulties in doing this is that everything I get has to be from a local store ...I'd like to please ask for help choosing a graphics card, motherboard and PSU, then RAM if necessary.

I'm not sure how we could help since we can't go to the store with you. Maybe you can go there and take notes or something?

 No.4113

>>4112
Sorry, to be clear I more meant I can't order off Newegg, Amazon etc or buy second hand because the purchase is being made by my mother's company and has to be done locally. It's a chain and they have a website: https://www.pbtech.co.nz/
Another store that sometimes has more niche products is https://computerlounge.co.nz/. I don't know if I can buy from them in this instance however.
The 9800 sounds good. I'm sometimes a bit wary of buying a very popular part, or even something popular outside of hardware, because it often seems like it's for a common use case, but I suppose I fit such a case when it comes to PC use. And thank you for reading my post, sorry I got a bit rambly.

 No.4114

>>4113
Hmm... Well, I'm not too familiar with AMD cards but I'd say you want at LEAST 12gb of VRAM. Personally I don't think you should bother buying a card with less than 16gb. If you can't afford the extra $100 New Zealand bucks (or the company rather) then save until you can.
But... hmm.. I just really don't know AMD stuff. I haven't had an AMD GPU since like 2002 or something. I'm just too paranoid over driver issues with obscure emulators and stuff. It's better these days, though.

 No.4196

>>2269
get a radeon for the same chipset at the same price chances are it's gonna be much better

oh wait january post. how are you doing anon? enjoying your new card?

 No.4197

>>4109
both 5070 and 9070xt are such a weird things. they are 5 years more recent then radeon 6800xt and still losing to it in many categories.

actually, speaking of weird, lets start how fucking bizzare it is that 5060 and 5070 are base 8 gb vram when 3060 had 12 lol

 No.4215

File:141131_OG_00001.mp4 (2.7 MB,600x720)

>>4196
>how are you doing anon? enjoying your new card?
Doing good! With AI I'm using my graphics card far, far more than I did when they were just for gaming. It used to be I'd get a decent card and then use it like 20 days of the year with specific games, but now I'm using it practically every day. I made a new SDXL NoobAI super merge to meet my demands and it took far less time since I was able to generate the images in a fraction of the time. I'm currently trying to hook up SillyTavern to SD so I can automatically generate images for scenes while the text AI is offloaded to OpenRouter. AI Video is kind of fun, but it doesn't hold the 'style' so it's not as good as I would hope. Unfortunately training video LORAs seems outside the realm even of the 5090 so I'm not sure when that would improve. People generally seem focused on porn, which is all well and good, but it quickly reverts to a generic appearance which kills the purpose of having 2D AI at all.

The only problem is AI makes it hard to find motivation to exist in real life or to create things, but I'm trying. The good news is that this card allows me to have blender open doing simple stuff unrelated to rendering, so I can AI prompt in the background while working towards other goals... slowly.

 No.4216

>>4215
what is the advantage of having your own ai model over using something like novelai or even chatgpt?

 No.4217

File:firefox_HyOYCeaZ31.png (45.11 KB,1285x833)

>>4216
Well, local models are better than online for the fact that you can customize them by adding LORAs and stuff, most local people do that. ChatGPT is of no interest to me because it's censored corporate stuff I can't use for escapism.
But when it comes to my obsessive model merging...
The model I made is a merge of like 8 other NoobAI-based merges. The advantage is selecting things I like and trying to optimize things. Some models are great at "creativity" (random mutations/errors) while others are good at poses or backgrounds or tag recognition and so on. By spending a bunch of time selectively merging different layers to try and isolate the best parts of each I try to maximize everything without sacrificing other things. For example layer OUT00 here has a huge influence on body types, so maybe someone else's model does great here, but layer OUT01 which controls the general coloring and shading is terrible, so I can take Layer OUT01 from the former and inject it into another model without the terrible colors.

 No.4218

>>4217
err I mean OUT00 is injected. Whatever.

 No.4220

>>4217
interesting. I use AI for human-level analysis. I mostly just use notebooklm, and feed it people's books, chatlogs, personal correspondence and any other source I can find, and use AI to draw this person's psychological profile, or look for inconsistencies in their narratives etc. Do you think I could benefit from having my own model?

 No.4221

>>4220
You're talking about text LLMs so the investment is like $10,000 to get a computer server rack setup thing that has like 800gb of fast RAM to load the entirety of the deepseek model, which has been the premiere local model for 2025 so far. Most people doing online stuff use the same model, but hosted by a company that largely charges nothing to use it. (I think they log the prompts for future training data)
The online-only models like GPT and Claude boast far larger context windows and tool use/agents so unfortunately the online stuff has a noticeable gulf in usability.

 No.4222

>>4221
Oh, but note that you can of course load smaller and weaker text models, but it's very much a hobby thing. People don't want to use something more prone to hallucinating when it comes to programming help and the roleplaying is far more dry since the data isn't there.

 No.4244

File:FusionX__00004.mp4 (10.82 MB,720x960)

AI Image Gen: Conquered. Well, hands are still an issue, but not as much. I'd wager I have the best 2D model in the known world for my specific needs.
AI Text Gen: Relatively great, I'm not rich so I can't do the best online models. I'm incorporating image gen and video emotion avatars.
AI Captioning: Yep. Got the NSFW tag thing going on.
AI Video: Making progress, but some weird errors keep popping up.

The 5090 has been very useful.
How can I make money from this while still being lazy...

 No.4246

>>4244
>How can I make money from this while still being lazy...
There’s some really obvious and cynical answers, but you probably don’t want to do them.

 No.4250

File:FusionX__00006.mp4 (9.81 MB,720x960)

>>4246
Yeah, probably. Stuff like spamming social media or actual art sites with an AI deluge until I get enough attention?
I don't want to make the world worse, I have my basic needs met so I don't need this to survive. Probably best to keep it as a tool to help me masturbate and design characters/scenarios for games.

Sadly this low sampler setup for video doesn't actually seem to work for me and I need to do the long generation times.
It seems like you can offload some VRAM blocks to CPU (or whatever the terminology) was, but since I can barely squeak by and fit this specific resolution inside 32gb of VRAM at 12 second run time, my generation time is pretty amazing. 5 minutes for this. It would probably be about 15 hours on the 3080 if I had, if I could do it at all.

 No.4436

I am now in the exact same situation and feel like my mind has already surrendered to making the worst financial decision of my life. I was originally gonna replace my 960 with a 3090 Ti, but now that I have seen just how much quicker the 5090 is for AI video alone, a feature I could actually really use, I know I wouldn't feel completely content with going for something lesser.
After doing some research, I could theoretically get one and pay in monthly installments of $100 at no interest, which coincidentally is the max. amount I usually save in a month, but I'd need to save like $50 on top of that for next year when my usual expenses increase in cost. I really need to find some loaded wallets on my weekly trip to the store..

 No.4437

>>4436
what do you need AI for? if you just play games, and used to be content with ancient shit like a 960
just buy radeon 6800xt or 7900xtx

Stay away from overpriced nvidia crap

 No.4438

File:[ASW] Ame to Kimi to - 02 ….jpg (209.59 KB,1920x1080)

>>4437
Not the same Anonymous but how is Radeon for AI? Been thinking of getting AMD GPUs for LLM/image/video gen use but most of youtube says AMD ones aren't as good as nvidia ones.

 No.4439

>>4438
AMD didn't try competing in that area until they released R9700 a week ago. Currently it's still not available for purchase. It is inferior to 5090, but ought to be cheaper.
Rather then watching youtube, use sites like nanoreview or technicalcity to compare the specs and look at benchmarks yourself. youtubers are paid shills.

 No.4440

File:FusionX__00018.mp4 (4.1 MB,720x960)

>>4436
OP here. I can say video was fun to mess with, but the time involved makes it more annoying than fun. Maybe it's just me, though. Video stuff is clearly made with realism in mind as it often struggles very much to maintain a 2D style if you tell it to do too many different things. Loops are the way to go since it's restricted to matching the end frame to the starting frame, reducing the freedom it has. If you're someone that finds 8 second loops of video far more erotic than still images then, yes, it's worth it. For me, though, it's usually too uncanny.
ACTUALLY NEVERMIND!
I forgot that a few weeks ago I spent like 20 hours messing with the ComfyUI nodes to combine two different workflows to try and make some good 2D looping. This actually looks pretty good. Well, definitely problems with her looking non-Kuon-y for a few frames, but that's why you'd generate like 20 videos and then go do something else. Alas, the randomness of AI is still here.

People keep making advances on making LORAs and similar things that allow you to generate faster, but... wait for it... it's made for real images.
Video models also use 'natural' language instead of booru tags, which I find really annoying, too. You can at least use other AI models to write the prompts or make suggestions. Video is definitely frontier territory, so keep that in mind. Also I want to say again how much I hate comfyUI.

For multitasking it's truly amazing. I can keep SD loaded and still play a game like Palworld and not feel anything, I can even have SD generating some images at the same time! Not video, though, video uses all the VRAM. If you want to do AI RP stuff with an on line text model you could also have some better stuff loaded to go along with it, like local voice model and stuff. All in all, the more you use AI the more you'll use the card.
It's kind of weird, now that graphic cards are also used for AI I use them about 5000% more often and now I actually feel like the cost is justified in a way.

 No.4441

File:FusionX__00023.mp4 (3.92 MB,720x960)

>>4438
If you're going to be doing AI stuff you need nvidia. There's no way around it. nvidia has CUDA, AMD has... nothing. nvidia's value wouldn't have become so extremely high if it wasn't a monopoly.

 No.4442

>>4438
AMD GPUs are much better if you're actually interested in driving a display(s) and having support 10 years from now. You don't have to rely on AMD to build drivers against the latest kernel and it is much less of a headache to keep it going.

Nvidia GPUs are effectively a black box. You're much better off buying them used and running them as second GPU. But even then you might not be able to continue using them forever. The only Nividia GPU I own now is not really that old but they no longer build drivers against recent kernels. So I'm stuck either running an old kernel or not using it at all. I hate having to plan my entire OS around the GPU.

I refuse to support them any longer because I've gotten burned too many times. There is really no point in using AI locally either. It isn't nearly as impressive as everyone is pretending it is.

 No.4443

File:[Serenae] Kimi to Idol Pre….jpg (209.23 KB,1920x1080)

>>4442
It's good to have principles and stuff, but the question is whether or not AMD is good for AI. The answer is a resounding 'no'. There isn't a choice. You pay the monopoly its entry fee or you're forced to sit out.

 No.4445

File:C-1753505545761.png (1 MB,1600x900)


 No.4446

>>4440
I’m fascinated by how you’re doing this. What’s your workflow like when making these generations? Is it at all hard to make it similar to what you’re picturing in your head?

 No.4449

>>4442
using AI locally is great if you have actually a reason to use it
if all you use it on is generating pictures of waifus and gifs like the nvidia shill does then yeah, it's just a luxury for the rich

also another argument: price needs to be considered. if you are serious about AI, you will need to invest. ideally at least 128GB of ram and multiple gpus. Here's the question of how much cheaper the R9700 will be once they start selling is separately. Projections are up to half the price of 5090. So here's the real question: what is better, a single 5090 or two R7900s? The answer to this question will be known once benchmarks start appearing. I'd wait a few weeks with buying a 5090 until then.

 No.4450

>>4449
>or two R7900

typo btw, meant two R9700s of course

 No.4451

File:FusionX__00042.mp4 (2.26 MB,720x960)

>>4446
Holy crap someone else is interested in this amazing technology?!
I only mess with image-to-video model as I can't imagine these are any good at generating 2D stuff from nothing. It's not going to know characters and stuff, so having a starting image is a huge requirement.
My basic process is:

1. Generate a good source image in SD
2. Use AI to do a basic natural language observation of the image + make suggestions on what to animate. I'm not sure if this is doing anything, I need to test it more. For example is "She moves her left hand smoothly to pick up the glass" really any better than "hand moves with glass"? I'm guessing it isn't.
3. Generate the video with a workflow thingie that took me a few days to actually understand.

Just like with image gen there's LORAs, but they're quite a hassle. Most of them are geared for realism as I said earlier so it's not that great, but some are really really nice for NSFW stuff.

 No.4454

File:[Erai-raws] Puniru wa Kawa….jpg (303.57 KB,1920x1080)

>>4449
>the nvidia shill
How much do you think nvidia is paying me to call their GPUs ridiculously expensive and to take months beating myself up over wanting to buy one?

You've entered my thread to call me a shill because I'm genuinely excited about a technology and sharing my personal experience on it to advise others. Not only that, but you're doing it purely out of ignorance.

Stable Diffusion/Image Generation? CUDA.
Audio Generation? CUDA.
Image Captioning? CUDA... I think? (most likely)
Video Generation? It took me a solid week to get the latest CUDA drivers to work with sage attention and pytorch.
Local Text Generation? This one is actually less dependant on CUDA. You can run text gen on CPU, but there's a large speed penalty. This is how people are running deepseek locally without spending $80,000.

Everything is built on CUDA and CUDA is nvidia.
Since you're proclaiming AMD over and over, do you know the name of AMD's CUDA analog?
I'll give you a second to think about it.


It's ZLUDA. It's not their own thing. It's a hack to try and get nvidia's CUDA to run on an AMD GPU and people have been working on it for years at this point. AMD as a company is not working towards it, just random people online. AMD forfeited.
If you think I have loyalty towards a monopolistic shit company like nvidia you are sorely mistaken. I would love it if there was actual competition in AI, but that is not our present world.

>what is better, a single 5090 or two R7900s?
For AI, the 5090. Unfortunately it's not even a comparison. AMD prices similarly to nvidia to such an extent that people said AMD blew it this year with its high prices, choosing to nickel and dime people instead of expanding their market share. They joined nvidia with the price gouging. Their "AI" card is being sold at a fraction of the nvidia price because it will be a fraction as useful.

 No.4455

>>4454
>You've entered my thread to call me a shill because
>my thread

no such thing, we are anonymous and this is a public place

 No.4456

>>4437
Aside from the obvious NSFW use, it has gigantic potential for 2D editing, think Eroge-style Motion Graphics that don't really follow a red line in terms of design. That's on top of other smaller time-saving benefits that add up over time, also related to workflow. I actually don't really play a lot of games myself.

>>4440
Thank you for your input, I was planning on asking!
What are your observations regarding the progression / optimization of local generation, does the community make an effort to try and make models less resource-intensive over time? Or is it similar to unoptimized AAA games with huge raw assets no one cares to compress, instead expecting the consumer to upgrade early? Because that's my main worry here. As someone who doesn't use camera footage, I know I will be set for non-AI editing for a decade+, but I haven't really been able to get a feeling regarding AI. For example, does image generation nowadays use less VRAM at more speed and equal / good results on, say, 20-series cards?

Also, what's up with the supposed lack of 2D-focused models? I think to remember that the threads on 4/h/ (might've been /trash/ instead, it's been a while) had something relatively exciting to report about every few days. To be fair, that was when the AI craze was still new and it was only related to images.

 No.4457

File:asdf.png (884.99 KB,1208x1093)

>>4456
>does the community make an effort to try and make models less resource-intensive over time?

Yes and no. Models themselves can't be made more efficient, it's basically a settled technology once it gets to the hobbyist's hands. It's the corporations and research teams that could make more efficient models, and then some rich guy could rent a data center to create a 2D model based on that tech. But he's not making a new technology, just training one based on another.

There have been a few efficiency improvements over the years in regards to VRAM usage and stuff, but I don't think much of it is new new, it's a few years old now. There's probably some gains for image gen dependant on ComfyUI, but I HATE ComfyUI and do everything I can outside of it. I know for video there's a VRAM offloader thing in ComfyUI that increases generation time but allows you to offload some stuff to RAM so you can generate the thing at all. I use it even though I have 32gb of VRAM. Video is THAT demanding once you get to certain sizes or lengths.

Anyway, those VRAM gains for image gen are offset by the newer models being made to generate larger images, which requires more VRAM. People also want to use the newer bells and whistles like upscaling and adetailer and controlnet and other stuff which each have their own VRAM cost. But you can definitely enjoy image gen with SDXL with 8GB or possibly even less if you heavily restrict active programs and limit yourself to maybe 1-2 images generated at once. I still have hardware acceleration disabled in all my browsers to free up a few hundred megs of VRAM.

There have been more improvements to LORAs/models that can generate things in fewer steps, which means less generation time which is most noticeable on people with weaker cards. However it generally comes at a cost somewhere, usually quality, and I never liked the ones I tried. (and the ones I didn't try require manual installation by altering code and that's above my brain). For video models, these faster generation things come at a cost not just of quality but they also tend to heavily restrict the animation in various ways, so each one I've tried left me disappointed. Except "lightx2" LORA, that one seems good and it's in all of my recent video things.

>Also, what's up with the supposed lack of 2D-focused models?
There's no money in it I would assume, or at least comparably. For video it's probably just that RL video data is far easier to come by when it comes to data scraping. And if somehow your dataset lacks something, grab a smartphone and record it yourself.

 No.4458

>>4457
Man, how disappointing about the excessive demands for video.. even if your 5 minute result from above does look very promising. What do you assume a 10s video would take to generate if you go for the maximum resolution that still fits into your VRAM with no offloading?

I stopped experimenting around the time people talked about LORAs, I just used the so-called "NAI leak" back then and decided that waiting 5 minutes for a single super low resolution image with horrible quality isn't worth the hassle. I've always wanted to get into training models with my favorite artists, but that's completely out of the question with my current card. I shall see if it's really gonna be the 5090, might just be the best "budget pick" on Nvidia when it comes to VRAM, though I haven't checked past workstation cards yet.

Either way, thank you very much for all the info! I shall make great use of it in the future.

 No.4459

File:firefox_IiI9iusojD.png (78.89 KB,1413x397)

>>4458
>Man, how disappointing about the excessive demands for video..
That's just for the extremes. 720x960 is quite huge for a video size and then the length greatly increases the VRAM usage. I can do 81 frames at 720x960 at like 22ish GB of VRAM.
My 'block offset' is 9 blocks (moved to RAM) when I do 161 frames, which is certainly on the extreme end of things. Videos tend to get errors with the colors at this length so I don't often try it. It might be a workflow thing, though.

>What do you assume a 10s video would take to generate
It goes by frames and not seconds, with adjustable framerate. 12 or so FPS seems to be about the slowest I can go without consciously noticing it's slow. 81 frames seems to be sweet spot, possibly what the model was trained at, so it would be a bit under 7 seconds.

 No.4460

>>4458
>What do you assume a 10s video would take to generate if you go for the maximum resolution that still fits into your VRAM with no offloading?

Oh and I'll time and test this later, but can't right now. It should be like 5-10 minutes when I push it to the limit of frames.

 No.4477

File:FusionX__00062.mp4 (2.91 MB,720x960)

>>4460
Okay, I did some testing. I'm using 12 frames a second.

720x960 resolution and 81 frames: 25gb of VRAM used and it completed in 171 seconds. I forgot that NOT offload blocks to RAM, but I assume the gains wouldn't be that major. Maybe a dozen seconds?
Don't know why it's behaving so weird with the colors. Such is video AI.

 No.4478

File:FusionX__00061.mp4 (3.98 MB,720x960)

>>4477
161 frames. 9 blocks offloaded to VRAM. 29gb used.
317 seconds. Maybe it's the pure white BG that's making it so weird.

 No.4481

>>4459
>>4460
>>4477
>>4478
Thanks so much again! It's nice to see how the first example seems to try being more creative with the animation. Most of the videos I have seen have this "animated PNG" feeling to them, similar to Gacha games. Have you read anything about RAM speed making a significant difference when offloading?

Also, would you mind sharing your sources when it comes to AI news that revolve around models designed for us 2D-focused people? Or really anything that would still benefit that genre.
I sometimes stumble upon random YT videos that showcase a large variety of things that seem incredibly interesting (and increase my 5090 itch), but it's just so much at once.

Last thing: We need more Koupen!

 No.4486

File:[Erai-raws] Puniru wa Kawa….jpg (281.68 KB,1920x1080)

>>4477
>>4478
Bleh, I now notice mistypes. I mean it's loaded from VRAM and not to it. The "offload blocks" thing moves it to RAM or something to reduce VRAM usage.

>>4481

>Also, would you mind sharing your sources when it comes to AI news that revolve around models designed for us 2D-focused people?
I just check civit.ai now and then to look at models directly. AI news, particularly youtube, is full of scammers and opportunists that promise you the moon so you'll click their thumbnail and are the same scum that did (and sometimes still do) the crypto hype stuff. I'll also rarely scan AI generals since it's easy to find informative posts as the guy will usually type more than 8 words, which makes it stand out from 95% of posts inside a general thread.

 No.4762

File:Screenshot 2025-08-31 at 0….png (299.36 KB,1645x761)

>>3893
>I feel immense guilt and shame over it
As you should! For ~$6K you can now get a DGX-1 system with 256GB of pooled VRAM. If you had saved your money instead of buying a 4090 and then a 5090 we could have an actual Kissu AI server. Tsk tsk tsk.

https://www.ebay.com/itm/167512048877

 No.4763

>>4762
I figure this is how AI stuff will go if you want to be serious about it. Buying a GPU and building a desktop around it isn't the best idea. If you get an Nvidia GPU now you'll have to deal with them dropping driver support in 5-10 years. This really sucks on Linux because you end up stuck on an old kernel or running the terrible FOSS drivers.

You're better off buying a GPU that provides full driver source code for your display devices and building a server for working with AI stuff. It's designed for that anyway. You send data off to be processed and it comes back.

Either that or wait until they start providing dedicated AI chips. I figure that is coming down the pike in the next few years. Built right into the motherboard (or at least a socket for it). I see us going to unified RAM soon instead of having the current RAM/VRAM split. Maybe this will improve the situation with buying a card to run displays and real time graphics. Right now things are so bad I've put off buying a new GPU for almost a decade. I can't justify spending what they're asking when my old one still does what I need.

I've moved most of the things I used to do on my desktop off to servers. All my files are stored there and it's running several things like a music+video daemon that can serve the data with metadata to whatever other device I want to use. Over the LAN and the internet. It has been much cheaper in the long run and having access to everything when I'm away from my main desktop has been very nice. It reduces how much I spend on other things like laptop and small handheld devices.

Prices are going up on the stuff you need for that to. But things like enterprise switches+routers+servers with server grade CPUs+RAM+HDDs are very cheap if you shop around. They dump so much of that on to the used market every year. The only thing that's really expensive still is the stuff you need to run a fiber network inside of your home instead of on CAT5/6 cabling. Which can push 10Gbps anyway and will probably get another speed increase with time.

 No.4771

File:[SubsPlease] Silent Witch ….jpg (462.98 KB,1920x1080)

>>4762
It does seem cool, possibly, but I'm beating myself up after scraping together $2800 so spending $6000 is simply impossible.
I'm really not sure how or if that thing you linked could be used for my AI purposes. From what I've seen having two GPUs instead of one complicates things, it's not just a plug and play kind of deal. You load THIS into GPU A, and THAT into GPU B, but you need to merge the result HERE and so on. I'm not some machine learning guy, I just run what other people have made.

The crux of the matter, though, is that this stuff is only as good as its adopters.
Right now there's been good progress in local video stuff because more people are getting 24gb and higher VRAM, which allows the hobbyists to create and share things. Without a community of people with the hardware it's not that useful.
As for text LLM stuff, 256GB is unfortunately still too low to load the best local models. There's a massive gulf and mid-range stuff largely doesn't exist.

 No.4919

>>4771
>still too low to load the best local models
256GB should be enough to run GLM-4.5-Air at FP16 (225GB), or GLM-4.5 at around INT4 (~190GB; Full-size is 755GB at FP16). Air has rough parity with GPT4o-mini, and GLM-4.5 rough parity with GPT4o. That's purely for knowledge-level tasks, though. No idea whether they're any good for RP or whatnot.

 No.4992

File:[Erai-raws] Busamen Gachi ….jpg (260.26 KB,1920x1080)

>>4919
Well, that's good to hear. That's still extremely out of my price range, though. Maybe we can start a Kissu Fund where everyone gives me thousands of dollars. What's the context window with that kind of setup?
That sounds really, really nice. Well, I guess it depends on the (E)RP potential, though. Ultimately people want the best of the best for things that matter, like programming, but ERP can be 'eh, good enough" if you're in the mood.

 No.4994

>>4992
>What's the context window with that kind of setup?
Probably in the range of 5-15K tokens... So... Usable with aggressive summarization, but maybe not so great for long, detail-oriented back and forths. Something something.... you can quantize the LLM, but not the context window (KV cache)?

Anyways, you can actually use GLM-4.5-Air and GLM-4.5 on https://chat.z.ai/. It's a bit faster than Deepseek, but nowhere near as fast as ChatGPT.

 No.4995

File:[Piyoko] Himitsu no AiPri ….jpg (262.92 KB,1920x1080)

>>4994
Are you still sending me the thousands of dollars?

 No.5088

File:Akane_seyana.png (28.13 KB,150x160)

>>4995
Sure, I'll just need the money upfront and I'll 10x whatever you give me.




[Return] [Top] [Catalog] [Post a Reply]
Delete Post [ ]

[ home / bans / all ] [ amv / jp ] [ maho ] [ f / ec ] [ qa / b / poll ] [ tv / bann ] [ toggle-new ]