Join devRant
Do all the things like
++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatar
Sign Up
Pipeless API
From the creators of devRant, Pipeless lets you power real-time personalized recommendations and activity feeds using a simple API
Learn More
Search - "gigabyte"
-
When i open a 2 gigabyte wordlist text file and forgot that my defaults text editor is GUI based.12
-
> Receive sudden phone call in the middle of the night
> Check caller, unknown number
> "Either something bad happened or it's something urgent. I'd better answer."
> "Hello?"
> Friend of friend of friend says he updated his gpu drivers and now has some random fps drops.
> I was in a good mood so I agreed to help him over teamviewer, even though I don't know him.
> Downgrading to an older version of nvidia driver seemed to have fixed the drops.
> 5 minutes later, he calls again. His headset is not working properly.
> Helped him fix the issue over teamviever.
> This continued for at least 2 hours, calling me every 5 minutes to install just another driver or change some random win setting. Turned out he had some retard format his pc because he thought it'd "make it go faster".
> Calls me again, this time he's pc isn't booting up at all. After 20 minutes on the phone the fucktard admits he just tried to reformat his pc because "my pc automatically installed a bad windows update" ( no, I don't understand either) but he fucked it up.
> I begin explaining him how to make a bootable usb stick, how to change the boot order etc to reinstall windows. I even suggested that I'd help him setup win/drivers after windows's done installing.
> He lets me go for about an hour explaining.
> "So that's it. When the setup is over, call me again and I'll help you install the drivers."
> "Bro this sounds complicated, why don't you come over? This won't even take you 5 minutes"
YOU MOTHERFUCKING PIECE OF SHIT
YOU FUCKING TRASH
CALL ME AGAIN AND I'LL SHOVE YOUR GIGABYTE GEFORCE GTX1060 6GB UP YOUR ASS, PERPENDICULARLY
The motherfucker even called me "bro"6 -
Noobies think there are 1,000 mb in a gigabyte. Real programmers know there are 1,024 meters in a kilometer.4
-
Modern software has gotten so bad that it even gets sluggish at times on late 2018 flagship devices. Slow, cheap hardware like is usually developers' and fanboys' excuse, particularly when it comes to Windows stuff? Like hell it is.
Software "engineering" has become so.. terribly inefficient. I'd dare any developer worth their salt to rewrite their program to make it work on an early 2000's machine. After all, those can run pretty advanced GUI's, have a reasonable amount of hardware (just think about how large a gigabyte of RAM really is) yet should be able to make for a reasonable limitation set.
Hardware limitations are the mother of optimization. Not every person on the planet has a 32-core Xeon workstation with 64GB of DDR4 RAM and a GTX Titan in it. Whether your application performs reasonably well on your machine shouldn't be the metric. Try deploying it on that laptop you tucked into a shelf years ago and reevaluate.. please.
And definitely you Slack!! Slacking off, is that what inspired the name of that pile of junk?! 😡26 -
GAHHHHHHHHHH! Damn USPS!!! Privatize them already! Stolen is more like it. It was a Gigabyte Aorus X470 Mobo, an AMD Ryzen 5 2600X CPU, and 16GB of memory. The postal service is a bloated, unresponsive, government run trainwreck. If people would just stop using them, they would shrivel up and blow away. ARRGH!!! 🔥☠️🔥☠️🔥23
-
Had 2 days of vacation. Theoretically (plus weekend, plus 2 days) 6 days.
Worked today… At Saturday.
Some administrators forgot to properly check bandwidth limitations....
*rolls eyes*
We had a major version upgrade of some server software at Monday.
Guess why I got called...
Of course it MUST be the software upgrade.
It couldn't be the new hardware that was setup 2 weeks ago and on which a lot of "important" VMs were migrated.
*eyes roll inside till only white is visible*
The even more annoying thing is that it wasn't that hard to figure out.
Looking at monitoring, we had spikes on 20 Gbit/s (roughly 2.x Gigabyte/sec - Ethernet) connection of some server at roughly 1.9 plus Gigabyte/sec.
IO latency spikes that made the graph look like a heartbeat EKG with severe tachycardia...
*additionally to white eyes starts cursing in reverse latin*
Incompetent admin answer: Booboo that can only be your fault - the developers must investigate.
Me (just a tad more polite): Meep Meep mother fucker, get your shit together. If the software would eat that much, the network would be a niece chunk of charcoal. Plus the time (sending instead of links to monitoring pictures… guess the lazy fucktard who's brain is a vacuum didn't even bother to check it)...
NOTICE SOMETHING?!
Incompetent admin: It starts at the same time. Always.
After wasting roughly another hour of time discussing with him, I just hanged up the video call.
Called someone I knew from the admin department and turns out that - drumrolls please - the incompetent admin was someone who got recruited 3 months ago…
*turning into antichrist*
I then had a not so polite discussion about how the only competent people could take days off (all except incompetent admin were on vacation) and the seemingly incompetent fresh recruit - who by the way NEVER mentioned this - was the only one left of the admin department. Which would be bad alone, but no - he even got the 24/7 emergency support role for the whole weekend.
Sometimes this company and HR especially notoriously drive me insane...
Guess next week there will be some HR barbecue.
But yeah. After a lot of raging around we nailed it down to the traffic of backups and could fix it.
Roughly 4 hours of analysis, communication, raging and hatred.
Just one hour implementing shit.
*goozfraba*11 -
I've optimised so many things in my time I can't remember most of them.
Most recently, something had to be the equivalent off `"literal" LIKE column` with a million rows to compare. It would take around a second average each literal to lookup for a service that needs to be high load and low latency. This isn't an easy case to optimise, many people would consider it impossible.
It took my a couple of hours to reverse engineer the data and implement a few hundred line implementation that would look it up in 1ms average with the worst possible case being very rare and not too distant from this.
In another case there was a lookup of arbitrary time spans that most people would not bother to cache because the input parameters are too short lived and variable to make a difference. I replaced the 50000+ line application acting as a middle man between the application and database with 500 lines of code that did the look up faster and was able to implement a reasonable caching strategy. This dropped resource consumption by a minimum of factor of ten at least. Misses were cheaper and it was able to cache most cases. It also involved modifying the client library in C to stop it unnecessarily wrapping primitives in objects to the high level language which was causing it to consume excessive amounts of memory when processing huge data streams.
Another system would download a huge data set for every point of sale constantly, then parse and apply it. It had to reflect changes quickly but would download the whole dataset each time containing hundreds of thousands of rows. I whipped up a system so that a single server (barring redundancy) would download it in a loop, parse it using C which was much faster than the traditional interpreted language, then use a custom data differential format, TCP data streaming protocol, binary serialisation and LZMA compression to pipe it down to points of sale. This protocol also used versioning for catchup and differential combination for additional reduction in size. It went from being 30 seconds to a few minutes behind to using able to keep up to with in a second of changes. It was also using so much bandwidth that it would reach the limit on ADSL connections then get throttled. I looked at the traffic stats after and it dropped from dozens of terabytes a month to around a gigabyte or so a month for several hundred machines. The drop in the graphs you'd think all the machines had been turned off as that's what it looked like. It could now happily run over GPRS or 56K.
I was working on a project with a lot of data and noticed these huge tables and horrible queries. The tables were all the results of queries. Someone wrote terrible SQL then to optimise it ran it in the background with all possible variable values then store the results of joins and aggregates into new tables. On top of those tables they wrote more SQL. I wrote some new queries and query generation that wiped out thousands of lines of code immediately and operated on the original tables taking things down from 30GB and rapidly climbing to a couple GB.
Another time a piece of mathematics had to generate all possible permutations and the existing solution was factorial. I worked out how to optimise it to run n*n which believe it or not made the world of difference. Went from hardly handling anything to handling anything thrown at it. It was nice trying to get people to "freeze the system now".
I build my own frontend systems (admittedly rushed) that do what angular/react/vue aim for but with higher (maximum) performance including an in memory data base to back the UI that had layered event driven indexes and could handle referential integrity (overlay on the database only revealing items with valid integrity) or reordering and reposition events very rapidly using a custom AVL tree. You could layer indexes over it (data inheritance) that could be partial and dynamic.
So many times have I optimised things on automatic just cleaning up code normally. Hundreds, thousands of optimisations. It's what makes my clock tick.4 -
New computer! I named it P.E.A.C.H. which is supposed to be an acronym, but I couldn't find a satisfying meaning.
Mind the stickers. I'd love to put a devRant sticker there too.
Components:
- Gigabyte AORUS GA-AX370-Gaming 5
- AMD Ryzen 7 1700
- Scythe Mugen 5 "PCGH Edition"
- 16GB HyperX FURY DDR4-2666 DIMM CL15-17-17
- 500 Watt Cooler Master B500 ver.2 Non-Modular 80+
- 275GB Crucial MX300 2.5" (SSD)
- 1000GB Seagate BarraCuda ST1000DM010 (HDD)
- Sharkoon BW9000-W Midi Tower
- Some video card and DVD drive I removed from an old computer
It still needs a good video card and internet connection.
With the dual boot and my MacBook Pro I cover Windows, Linux and MacOS.20 -
I finally finished building my desktop to the specs I want!
I bought NZXT's Nuka-Cola themed case & mobo cover when they were announced. I've been planning this build for a while, but:
My laptop fell off a bench (while in my backpack) a couple months ago and the screen broke, so I bought this nice CHG70 from Samsung and put on top of it. That worked fine for a while, but within a couple weeks that laptop also stopped sending a display signal.
Having already dropped a lot of money on the monitor of my dreams and not being able to bear returning it after having it in front of me, I decided, Fuck It; I'll just build the whole PC I've been planning on right now.
Except, I wasn't ready. Had to start out with a Ryzen 2400g. Then got an RX580 on sale for $200. This week was when I swapped the Gigabyte B350M DS3H & 2400g for NZXT N7 Z370 & i7 8700k18 -
Our team makes a software in Java and because of technical reasons we require 1GB of memory for the JVM (with the Xmx switch).
If you don't have enough free memory the app without any sign just exits because the JVM just couldn't bite big enough from the memory.
Many days later and you just stand there without a clue as to why the launcher does nothing.
Then you remember this constraint and start to close every memory heavy app you can think of. (I'm looking at you Chrome) No matter how important those spreadsheets or illustrator files. Congratulation you just freed up 4GB of memory, things should work now! WRONG!
But why you might ask. You see we are using 32-bit version of java because someone in upper management decided that it should run on any machine (even if we only test it on win 7 and high sierra) and 32 is smaller than 64 so it must be downwards compatible! we should use it! Yes, in 2019 we use 32-bit java because some lunatic might want to run our software on a Windows XP 32-bit OS. But why is this so much of a problem?
Well.. the 32-bit version of Java requires CONTIGUOUS FREE SPACE IN MEMORY TO EVEN START... AND WE ARE REQUESTING ONE GIGABYTE!!
So you can shove your swap and closed applications up your ass but I bet you that you won't get 1GB contiguous memory that way!
Now there will be a meeting about this issue and another related to the issues with 32-bit JVM tomorrow. The only problem is that this issue only occures if you used up most of your memory and then try to open our software. So upper management will probably deem this issue minor and won't allow us to upgrade to 64-bit... in 20fucking1910 -
I spent the last three weeks+ (literally THREE full weeks, weekends too) building something I thought was really cool, powerful, and useful. Made a blog post, posted a giant thread on the company Twitter.
Literally one person gave it a like.
I don't know why I give shit anymore, cuz nowadays if it isn't about getting rich quick, cHaTgPt, or some other made up hype, no one cares. Apparently I shouldn't either...
Meanwhile my 16 GIGABYTE RAM MAC, yes 16 GIGABYTE RAM can't even hold power while plugged in, and I'm still clowning around with an ancient iPhone 6 (actually one of my mom's old iphones) that barely stays above 20% battery for more than an hour...
And FINALLY, my FUCKING ISP is for sure screwing me, since I've been doing some hard core data streaming and broadcasting, even though I pay $60+ month for that shit it, keeps dropping out, shit doesn't load.... I mean wtf this isn't 1990 dialup AOL anymore
When I step back I just feel like the worlds biggest loser, maybe the world's biggest 🤡7 -
Buffer usage for simple file operation in python.
What the code "should" do, was using I think open or write a stream with a specific buffer size.
Buffer size should be specific, as it was a stream of a multiple gigabyte file over a direct interlink network connection.
Which should have speed things up tremendously, due to fewer syscalls and the machine having beefy resources for a large buffer.
So far the theory.
In practical, the devs made one very very very very very very very very stupid error.
They used dicts for configurations... With extremely bad naming.
configuration = {}
buffer_size = configuration.get("buffering", int(DEFAULT_BUFFERING))
You might immediately guess what has happened here.
DEFAULT_BUFFERING was set to true, evaluating to 1.
Yeah. Writing in 1 byte size chunks results in enormous speed deficiency, as the system is basically bombing itself with syscalls per nanoseconds.
Kinda obvious when you look at it in the raw pure form.
But I guess you can imagine how configuration actually looked....
Wild. Pretty wild. It was the main dict, hard coded, I think 200 entries plus and of course it looked like my toilet after having an spicy food evening and eating too much....
What's even worse is that none made the connection to the buffer size.
This simple and trivial thing entertained us for 2-3 weeks because *drumrolls please* none of the devs tested with large files.
So as usual there was the deployment and then "the sudden miraculous it works totally slow, must be admin / it fault" game.
At some time it landed then on my desk as pretty much everyone who had to deal with it was confused and angry, for understandable reasons (blame game).
It took me and the admin / devs then a few days to track it down, as we really started at the entirely wrong end of the problem, the network...
So much joy for such a stupid thing.18 -
Jesus fuck Gigabyte motherboards downloading and installing firmware updates over HTTP no fucking S
https://tomshardware.com/news/...10 -
Me: I've got this wonderful idea, I'll try it in swift
OSX: download the 4.47 GIGABYTE Xcode
Me: :'( ok
Xcode : you need to update your OS and that's again more than 4.5 GIGABYTE
Me : Sublime -> Main.class1 -
NewLifeNewHope update No.2 / Day 5
My Server Is finally opened and i installed 2x8 TB WDC Harddrives and SATA 128 SSD. I know this server is freaking garbage, but i got this PC for free, and have somewhat good-ish upgrade path, so heres the spec :
-Gigabyte H110m-S2 LGA1151
-Intel i3-6300
-8GB DDR4 single channel RAM
-128 GB SSD
-2x8TB Harddrives
-TP-Link 1000mbps NIC
so the plan is to make this server as the Main Repository -- yes no offsite backup plan for now -- and also i want to make this server as an Email Server and for hosting my company's website. I've already asked for static IP from my ISP and will take effect tommorow.
I need help for choosing the Operating System (i like centOS) for my server. and to setting this Server to work like what i planned but i don't know where to start, Any help and/or References will be great !14 -
MTP is complete garbage. I want mass storage back.
The media transfer protocol (MTP) occasionally discovers new creative ways of failure. Frequently, directory listings take minutes to load or fail to load at all, and it freezes up infinitely (until disconnected) when renaming an item, and I can not even do two things simultaneously.
While files are being moved, I can not browse pictures or watch videos from the smartphone.
Sometimes, files are listed with the date 1970-01-01 (Unix epoch) instead of their correct date. Sometimes, files do not appear at all, which makes it unsafe to move directories from the device.
MTP lacks random access. If I want to play a two-gigabyte 4K 2160p video and seek in the video, guess what: I need to copy it to my computer's local mass storage first because MTP lacks random access.
When transferring high numbers of files, MTP has to slooooowly enumerate (or "prepare" or "calculate the time of") them all, which might even take longer than mass storage would need for the entire process. This means MTP might start copying or moving the actual files when mass storage is already finished.
Today, the "preparing to move" process was especially slow: five minutes for around 150 files! How am I supposed to find out what caused this random malfunction?
MTP sometimes drives me insane. I want mass storage back, at least for the MicroSD memory card, which uses a widely supported file system.
Imagine a 2010 $100 Android phone is better at file transfer than a 2022 $1000 Android phone (or iPhone, for that matter).3 -
One thing that really bothers me is that people use use megabyte to mean mebibyte. What makes it even worse is that Windows does it too. It says it shows gigabyte but actually shows gibibyte.
Now imagined my frustration when I saw this D:13 -
GOD DAMN IT COLLEGE YOU DID IT AGAIN. for real college can go suck Satan's 50 inch red cock for all I care.
A professor asked me to design a processor and I'll get a bonus. I said okay cool nothing hard.
oh but it has to be in verilog.
okay cool.
oh and it has to be on this fucking ancient useless piece of shit called xilinx that the fucking college provides to you only via a fucking 50 gigabyte virtual machine.
sigh. okay..... challenge accepted.
It fucking crashes every 2 minuites. And after 3 days of no sleep. I finally finished the Alu, Control unit, 4k memory, 8 registers and the busses.......... BUT THEN THE ENTIRE VIRTUAL MACHINE CRASHED AND LOST ALL PROGRESS...... fml.
and the professor only gave me the bonus for the Alu. sigh. fuck college.11 -
Least successful...
In a nutshell, an multi version http client for a elasticsearch.
It supported ES 1.7 up to v7.
With an reduced future set, but all in all it allowed doing everything ES offered - just not for one version, rather the whole monty.
For various reasons I wasn't allowed to opensource that...
Which brings me to the least successful part. The client is a beast and would be a blessing for a lot of people I'd guess, but it's sadly covered by more legalese than one could imagine.
Think of legalese as in "Angel - Wolfram and Heart" legalese. I wouldn't be surprised if some part of the contract was written in blood.
... And least successful as in: Nope. Never gonna do that again.
Abstractions necessary for supporting multiple versions are are really painful.
Having an E2E test suite consuming > 64 Gigabyte of RAM for testing against several ES docker instances in parallel isn't fun.
Nothing of that project was fun.
Still gives me nightmares.
(NDA expired short time ago) -
Finally able to complete my almost-aorus build!
Ryzen 5 3600
Gigabyte Aorus B450 Pro Wifi
Team Delta 2x16gb 3600mhz
Western Digital 2tb 3.5
Crucial P2 500gb NVME m.2
Gigabyte Aorus RX 5700 XT 8gb
Aorus GP AC300G Tempered Case
Gigabyte Aorus P850w 80 gold
Gigabyte Aorus Cooler 360
ID Cooling XF 12025 RGB Trio
Viewsonic XG2405 24in IPS 144hz 1ms
Cables are not managed yet, sorry. And I like listening to Blankpink 😋20 -
When file managers copy and delete files within the same partition instead of moving or renaming them…
When Google's Storage Access Framework was introduced, it did not feature a move command, so file managers just resorted to copying and deleting files within the same storage. Not only does this cause needless wear and is much slower, but it also destroys the date/time attribute (it gets changed to current).
When moving files through MTP (miserable transfer protocol, used for connecting smartphones to PC), they are also copy-deleted. This makes moving a 20-Gigabyte DCIM folder impractical. Also, if one cancels the operation, it might end up whoopsie-daisy deleting some files from the source before they have been transferred.
MTP is so bogus that it is incapable of a simple operation that would JustWork™ on mass storage devices. Not to mention, MTP lacks parallelism and its directory listing loading it S-L-O-W. Upwards of a minute for just 1000 files. Sometimes, it fails loading at all.
Also, trying to rename a file through MTP using the terminal through GVFS, even if just within the same folder, it copy-deletes it. If I want to rename a 1 GB 2160p 4K video in a highly populated DCIM folder, I can not do so through the terminal. At least, the 4K video has a time stamp in its internal metadata, but it still renames slowly and adds needless wear to the smartphone's flash memory.14 -
If you publish a package over Composer - or in fact any package manager that allows optional dependencies - please don't make dependencies that aren't **absolutely necessary** required.
Is it a theme for your nice app? Don't make it required, some are not going to want to use this theme and don't want to deploy extra Megabytes to their servers every single time.
In fact, someone may even want to replace that dependency with a fork, like a customized version of that theme because it is not flexible enough! Forcing that dependency means it can only be replaced with hacks.
I am looking at you, `oxid-esales/oxideshop-metapackage-ce`, with your extra themes and demoshop data that makes the Deployer tool push and copy around half a Gigabyte of unused assets on a website every single deployment!!2 -
I finally got my stickers. I feel like I have transcended existence. Or maybe successfully cleared a merge conflict! 😂
-
What is a good ultrabook laptop for about $1600, more or less?
I am currently looking at the Gigabyte Aero 14 and the Dell XPS 15
Any advice will be welcome3 -
So Im planning to build a pc, which i will mainly use it for dev and gaming in free time, my main components will be:
CPU: INTEL 8700K
GPU: GTX 1080 msi or gigabyte?
SSD: 860 EVO
RAM: 16GB 3200MHZ
MOTHERBOARD: should i go with msi or gigabyte whixh one is better?
PSU: 650W or 700W deepcooler?
Also for the cpu cooler do i get water colling or a standard cpu fan?
P.S: i plan to overclock the cpu and gpu at some point.
Also whats your opinion on the rgb lightning gpu and motherboard, and is there point in getting a mobo with sli support (is it work buying second gpu at some point or better upgrade the existing)4 -
Quess who's back again, php oudated piece of shit monolith codebase. So we have a relatively huge client we need to migrate to AWS.
It is written with yii, all object-oriented. The way it's implemented makes me question my love for object oriented as well my sanity for even accepting this project.
I probably could talk about this piece of shit for hours but the fact they save 3 gigabyte of qr code images is the fucking worst. It's literally a few one hundred thousand images who could be generated on the fly.
Please for the love of god, let me finish this migration tomorrow.4 -
On Dailymotion, failed uploads count towards the 24-hour rate limit.
Dailymotion has a rate limit of somewhere between 10 to 15 videos (appears to vary). I experienced a glitch where I dragged 10 videos into the uploader (the highest number; years earlier it oddly was 22), and none of the uploads would start. However, it still counted towards the daily rate limit, immediately blocking me from uploading for 24 hours. I have a slight suspicion that this failure was deliberate.
Also, that rate limit is indiscriminate of video size. A gigabyte-sized 4K video counts equally towards the rate limit as a 7 MB 240p video.1 -
No internet! I have been without home WiFi for a week now and I can't download big files over my hotspot as it is expensive. I need to download a Raspbian image for my Raspberry Pi but it is ~1 gigabyte. Been needing to get my Raspberry Pi set up to turn it into a server for a site I am currently creating.6
-
Installing Linux distros on ryzen gigabyte motherboard was a pain in the ass. I gave up on Ubuntu and mintlinux and went onto install fedora with malfunctioning mouse(driver issue, mouse was fine). Basically had to hack around with only keyboard for few hours before I could install the latest Nvidia drivers and got my mouse to work. I don't who is at fault here but the Linux community needs to put enough pressure on vendors like gigabyte/amd to make sure they release updates for their hardwares across os. Installing an Os shouldn't be this hard
-
spent a few days trying to track down the cause of a thermal shutdown in my workstation. intel 4790k with no overclock would spike to 95C on one core (core1) whenever maxing out all 8 threads, be it real work, mprime, anything with 100% cpu being used. I quadrupled my RAM from 8gb to 32, because its cheap and id like to have all data in memory sometimes, not because I thought that was the problem. I reseated my watercooling block. I checked out the PSU. I unplugged all unnecessary peripherals, drives, etc. It turned out to be a bug in Gigabyte MOBO BIOS (causing temps to be read incorrectly i think, still not exactly sure...) updated from version 5 to 10 and poof now temps are back in the high 50's at full load. it only took 2 days to figure out and i think i learned something
-
Been working on redoing my desktop lately. Currently the specs are:
-FX-8350
-Gigabyte GA-970A-DS3P motherboard with a broken USB 3.0 header lmao
-GTX 660 (Gonna upgrade to an RX 580 at some point, I don't do any hardcore gaming so I know I don't need a top of the line GPU)
-Crucial BX500 240gb SSD
-WD 500gb HDD (gonna upgrade to a bigger one eventually)
-Some like $60 Dynex PSU I bought a while ago, waiting on my Corsair RM650x to come in
At this very moment, it's running Windows 7 Ultimate x64. Once I get to a point where I'm happier with the build, I'll switch it over to Linux and start ricing. It has Windows right now cause I'm just using it for some games and when I last fucked with the hardware, it was the middle of the night so I didn't want to spend too much time setting up a Linux distro the way I want it and everything right then, just putting that off for later (especially cause I use Arch btw)
I have been playing some Half Life 2 lately. I forgot how fucking fun that game is.
Aside from my PC, my birthday was technically yesterday (it's about 2:30AM as of writing this, and I've been up for a while, so I still consider it today). Now I'm 2 years away from being able to legally drink (and smoke since the law change, although I still do both anyways).
I'm gonna stop rambling. Life is fairly decent right now. Not too much to "rant" about except for shit with my roommates, but I won't bore everyone with that1 -
Building a dev/gaming PC - What do you guys thing about these specs:
+ Mouse Cooler Master Storm Mizar SGM-4005
+ Intel i5-6600K
+ SSD 2.5" Samsung 850 Evo
DIMM 8GB DDR3 1866Mhz Kingston CL10
+ Cooler Master Hyper 212
+ PSU 750W CoolerMaster GM series G750M
+ PCX GeForce GTX 1070 8GB GDDR5 DVI/HDMI/3xDP VR Ready DX12
+ Monitor 25" Dell UltraSharp U2515H LED IPS, 16:9, QHD 2560 x 1440
+ Keyboard Gigabyte K85 Mechanical Gaming
+ ATX Case Fractal Design Core 3500 Windowed
+ Motherboard (not sure yet)8 -
Does anyone familiar with overclocking Ryzen non-x cpu? I had a Ryzen 7 2700 with gigabyte B450 motherboard. The core voltage locks to 0.9v after clock ratio set to 34 from auto (3.2GHz). But on auto, the vcore is around 1.38v, which is close to max recommendation. Why?9