Join devRant
Do all the things like
++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatar
Sign Up
Pipeless API
From the creators of devRant, Pipeless lets you power real-time personalized recommendations and activity feeds using a simple API
Learn More
Search - "graphics issues"
-
Pm: "so how long should this take?"
Me: "well, I see at LEAST 8 hours of work here for me, some of which can't be done without graphics, but let's assume graphics are done by designers by the time I get to hour 6 and I just have to plug the images in without much editing or tweaking."
Pm: "So you'll have this by tomorrow?"
Me: "Do you think I just sit on my hands all day waiting for one of you trolls to give me something to do? I said AT LEAST 8 FULL HOURS OF MY TIME, that would be assuming I could get 8 consecutive hours to work on this, and let's face it that's not gonna happen."
Pm: "So Friday...?"
Me: "Tell the client that so long as there aren't any unforseen issues it could be done as early as Thursday. I'll put it on my schedule, but it is not a priority until at least Monday. There's only one of me and 5 other active projects on my desk."
Pm: "I'm gonna tell them Wednesday morning we'll show it to them for approval."
What in the hell is wrong with people?4 -
So, today for my SO's father who is already over 70 and wants to try Linux. However, he doesn't want Linux on his main PC for now, rather on the old one so that he can take his time to get familiar, which is a reasonable plan.
But holy crap, what a machine! Intel Core2 Duo 4400, 2 GB DDR2(!) RAM, 250 GB IDE(!) HDD, DVD RW drive. Graphics, sound and LAN integrated on the mobo chipset. It's half a miracle that it doesn't run on steam. The machine had been delivered with Vista and has always been painfully slow.
It doesn't even support booting from USB, but I had prepared a DVD just in case. Surprise: it booted from DVD without issues and with full HW support!
Partitioned and installed, deleted Vista in the process (felt good). I went with the full blown Mint 20 Cinnamon edition because XFCE isn't as beautiful. Also, having XFCE now and then Cinnamon looking different on the other PC would be confusing.
Installation took some time, but worked. Cinnamon's RAM usage is at 750 MB idle, and at 1.1 GB with Firefox started. Once the PC is booted, it runs pretty OK with reduced swappiness and noatime on all file systems, plus unnecessary startup applications disabled. Updates took long, but ran through successfully. Installed LibreOffice and some small games, Firefox got uBlock Origin, Youtube worked OOTB.
That PC somehow had escaped disposal several times - and now has a proper OS for the first time in its miserable existence. It runs so much better than it ever has. Just wow, a "big" Linux desktop from 2020 blows a contemporary Vista out of the water on such an old machine!16 -
I've run into my first nVidia driver issues. Ever. ☹
I'm attempting to install the drivers on my shiny new Dell XPS15 8750. It has both an Intel (default) and nVidia graphics card; the Intel works just fine out of the box with Nouveau, with terrible fps. Installing the nVidia drivers causes a conflict with those, and X just displays a black screen. I can still interact via the virtual terminals, but with a 4k display it's kind of annoying.
I'm currently trying to install nvidia-bumblebee and related crap, bur it is not going well.15 -
A friend came to me whether i want to do a project on c++(someone asked him to find a c++ guy).
Me needing money didn’t refuse. Even though i am a Java developer with 0 skills on c++, but wanted to give it a try.
So project started, and it was about a plugin for rhinoceros app(3d graphics app).
The plugin was simple, had some views and some services to upload a file into s3 and some api calls, not something complex..
So i ended up working on the project together with my friend(web dev).
So long story short, we had a lot of issues, but considering we both had no knowledge on c++, we were really lucky to finish the product almost on time(3 days after).
Did no memory management even though i’ve read that we have to do that by our selfs and that c++ doesn’t have garbage collector.
But the plugin worked great even without garbage collector.
Had a lot issues with string manipulation, which almost drive me crazy.
PS: did a post here before taking the project, to ask whether it is a good idea to take the project or not, had some positive and some negative replies, but i deleted the post since i thought i was breaking the NDA i signed 😂😂
PS2: just finished OCAJP 8 last week with a great score😃6 -
So recently I had an argument with gamers on memory required in a graphics card. The guy suggested 8GB model of.. idk I forgot the model of GPU already, some Nvidia crap.
I argued on that, well why does memory size matter so much? I know that it takes bandwidth to generate and store a frame, and I know how much size and bandwidth that is. It's a fairly simple calculation - you take your horizontal and vertical resolution (e.g. 2560x1080 which I'll go with for the rest of the rant) times the amount of subpixels (so red, green and blue) times the amount of bit depth (i.e. the amount of values you can set the subpixel/color brightness to, usually 8 bits i.e. 0-255).
The calculation would thus look like this.
2560*1080*3*8 = the resulting size in bits. You can omit the last 8 to get the size in bytes, but only for an 8-bit display.
The resulting number you get is exactly 8100 KiB or roughly 8MB to store a frame. There is no more to storing a frame than that. Your GPU renders the frame (might need some memory for that but not 1000x the amount of the frame itself, that's ridiculous), stores it into a memory area known as a framebuffer, for the display to eventually actually take it to put it on the screen.
Assuming that the refresh rate for the display is 60Hz, and that you didn't overbuild your graphics card to display a bazillion lost frames for that, you need to display 60 frames a second at 8MB each. Now that is significant. You need 8x60MB/s for that, which is 480MB/s. For higher framerate (that's hopefully coupled with a display capable of driving that) you need higher bandwidth, and for higher resolution and/or higher bit depth, you'd need more memory to fit your frame. But it's not a lot, certainly not 8GB of video memory.
Question time for gamers: suppose you run your fancy game from an iGPU in a laptop or whatever, with 8GB of memory in that system you're resorting to running off the filthy iGPU from. Are you actually using all that shared general-purpose RAM for frames and "there's more to it" juicy game data? Where does the rest of the operating system's memory fit in such a case? Ahhh.. yeah it doesn't. The iGPU magically doesn't use all that 8GB memory you've just told me that the dGPU totally needs.
I compared it to displaying regular frames, yes. After all that's what a game mostly is, a lot of potentially rapidly changing frames. I took the entire bandwidth and size of any unique frame into account, whereas the display of regular system tasks *could* potentially get away with less, since most of the frame is unchanging most of the time. I did not make that assumption. And rapidly changing frames is also why the bitrate on e.g. screen recordings matters so much. Lower bitrate means that you will be compromising quality in rapidly changing scenes. I've been bit by that before. For those cases it's better to have a huge source file recorded at a bitrate that allows for all these rapidly changing frames, then reduce the final size in post-processing.
I've even proven that driving a 2560x1080 display doesn't take oodles of memory because I actually set the timings for such a display in order for a Raspberry Pi to be able to drive it at that resolution. Conveniently the memory split for the overall system and the GPU respectively is also tunable, and the total shared memory is a relatively meager 1GB. I used to set it at 256MB because just like the aforementioned gamers, I thought that a display would require that much memory. After running into issues that were driver-related (seems like the VideoCore driver in Raspbian buster is kinda fuckulated atm, while it works fine in stretch) I ended up tweaking that a bit, to see what ended up working. 64MB memory to drive a 2560x1080 display? You got it! Because a single frame is only 8MB in size, and 64MB of video memory can easily fit that and a few spares just in case.
I must've sucked all that data out of my ass though, I've only seen people build GPU's out of discrete components and went down to the realms of manually setting display timings.
Interesting build log / documentary style video on building a GPU on your own: https://youtube.com/watch/...
Have fun!20 -
Holy fuck EA (yea I went there)
So here I am playing NFS Heat, and my laptop is hot enough to cook eggs on the keyboard, no joke, yet spec wise I should have 0 issues running this game.
I switch on the turbo cool (basically spin the internal fans at full speed... It's an MSI) and this usually covers any heating issues, nope... not this time.
this fucker of a fan can't keep up , so I've turned down the settings from ULTRA, that's depressing, to Medium... it feels like 2006 showed back up, but here's this laptop still burning a whole in my fingers.
My GPU (GTX1060 6GB) is running at 95'c on Medium graphics, and LOW looks like I fell into a 8bit world with slow ass rendering.
Guess I'm going to need to get some cooling assistance for this thing.
Thank you EA for turning my 2.5k laptop into a stove top, I appreciate it.
I guess the name checks out, "heat"14 -
I'm not sure if this should be a joke or random, but the SBB train info screen in St. Gallen on platform 5 just decided to interpret the data given a bit differently than intended.
It is not broken and no other screen has that problem. What could it be, I wonder. But I don't care enouht to post it as a question.
I guess random would have been more appropriate.13 -
Ended up with 16.04 and 18.04 ubuntu in dual boot. (see last rant).
Fixed graphics issues on 18.04. The fix envolved setting stuff in grub. Ok.
Now we apparently need 16.04 again and can kich the 18.04.
np. Let me just boot my 16.04
oh look who has graphics issues now!
It boiled down to the grub settings. so Depending on if I need 18.04 or 16.04 I need to fiddle with grub. great.
And now we don't neet 18.04 at all. So I tried to do something not in the last minute once, and it backfired completly. Wasted 5h on this shit this week. Fk this.2 -
!rant, need advice
I have no idea where to start with freelancing.
My girlfriend's been wanting to get some stuff but due to some issues hasn't been able to and has been extremely hesitant about bringing the topic up.
So to surprise her, I decided it'd be nice if I could freelance on the side and rake up some money so she can get the things she wants and I can spoil her (she's been through things and honestly deserves it).
So if you guys have any suggestions on where to start freelancing, that'd be great. I can do graphics and web design and just set up an anonymous Fiverr account but never had any luck with it.6 -
Me: ugh I really hate the lack of graphics performance in virtual box and not a fan of how CPU hungry VMware can get... Fuck it let's try gnome boxes
Also me: *spends 4 hours diagnosing issues with gnome boxes and Ubuntu 18.04 based distros*
But so far... Worth it. Seems pretty damn good so far1 -
Kudos to Fedora for the 32 workstation release. It works as advertised on Lenovo laptops (even the new ones) OOTB, no issues.
Caveats: It's Wayland with Nvidia graphics disabled and has no fingerprint support, but that goes with the territory. Still an arch adherent personally, but it makes the office engineer laptop story significantly easier. -
so my poor old hp bro is starting to give up on me after 3 years . there are both battery and performance issues. I knew its death was near when i got its keyboard replaced i saw its battery's state 6 months ago. it was fat as me . Now it can't live without the wire for more than 30 mins.
apart from that am getting frustrated with everyday performance of my system. all i open is 50 tabs of chrome and android studio where bitch gradle generates a billion byte sized files at million places and cache folders . (and sometimes a terminal and/or files and/or vscode).
I have decent specs tho i5 7th generation/8gb ddr4 ram/ 2gb nvidea graphics/ 1tb hdd. but still my ubuntu gets stuck everytime i switch between chrome and AS . (maybe i have not made correct swap partitions or maybe there is an issue due to ubuntu/windows dualboot)
what should i do ? I guess i have to spend some shit on a new battery. But apart from that, iwanted to know about performance. how to get a beter performance?
I have heard of solutions like getting an sdd or increasing ram, but am broke af and might not afford a 1tb ssd (yes i do need that much amount of memory, my system is currently at 650 something gb) and about ram i have heard it doesn't offer much improvements. is that true?What should i do?6 -
!rant
You know, for the first time, I have had issues diagnosing an error (due to not having slept in about 75 hours due to the CEO slamming down on me to finish a project), and I posted on a forum. First time doing that, and they are fucking useless.
I may as well have delved into google and tries to diagnose the error. Done so, but you now what I find? People with the same fucking error, but no fucking resolution.
Stupid issue is a desktop support one, bloody Ubuntu won't install (gives me a shitty error and then loops back to installation post restart of my laptop. More than sure it has something to do with my graphics card).
Laptop decided to crash at the last moment, hence me installing Ubuntu again. Fucking errors, fucking help forums not helping, and fucking tiredness. On the plus side, coffee has kept me going longer than I thought. -
Design in Motion: Real-Time Rendering's Impact on Architecture
Architecture, a discipline that once relied heavily on blueprints, models, and lengthy render times, has undergone a revolutionary transformation in recent years. The advent of real-time rendering technology has fundamentally altered the way architects visualize, present, and interact with their designs. This paradigm shift has not only enhanced the creative process but has also empowered architects to make more informed decisions and create immersive experiences for clients and stakeholders.
Real-time rendering, a technological marvel that harnesses the power of high-performance graphics hardware and advanced software algorithms, allows architects to generate photorealistic visualizations of their designs in a matter of milliseconds. Gone are the days of waiting hours or even days for a single rendering to complete. This acceleration in rendering time has not only expedited the design process but has also encouraged architects to explore multiple design iterations rapidly.
One of the most significant impacts of real-time rendering on architecture is the ability to visualize a design in various lighting conditions and environmental settings. Architects can now instantly switch between daytime and nighttime lighting scenarios, experiment with different materials, and observe how their designs respond to different seasons or weather conditions. This level of dynamic visualization offers insights into how a building's appearance and functionality evolve throughout the day, contributing to more holistic and thoughtful design solutions.
Moreover, real-time rendering has transformed client presentations. Architectural concepts can now be communicated with unprecedented clarity and realism. Clients can virtually walk through spaces, observing intricate details, exploring different angles, and even experiencing the play of light and shadow in real-time. This immersive experience fosters a deeper understanding of the design intent, enabling clients to provide more targeted feedback and make informed decisions.
The impact of real-time rendering on collaboration within architectural teams cannot be overstated. Traditionally, architects and designers would need to wait for a rendering to complete before discussing design changes or improvements. With real-time rendering, team members can make adjustments on the fly, observing the immediate effects of their decisions. This seamless collaboration not only enhances efficiency but also encourages interdisciplinary collaboration as architects, engineers, and other stakeholders can work together in real-time to refine designs.
The integration of virtual reality (VR) and augmented reality (AR) into the architectural workflow is another transformative aspect of real-time rendering. Architects can now create VR environments that allow clients to step inside their designs and explore every nook and cranny. This not only enhances client engagement but also enables architects to identify potential design flaws or spatial issues that might not be apparent in 2D drawings. AR, on the other hand, overlays digital information onto the physical world, facilitating on-site decision-making and construction supervision.
Real-time rendering's impact extends beyond the design phase. It has proven to be a valuable tool for public engagement and community involvement in architectural projects. By creating virtual walkthroughs of proposed structures, architects can offer the public an opportunity to experience the design before construction begins. This transparency fosters a sense of ownership and allows for constructive feedback, contributing to the development of designs that resonate with the community's needs and aspirations.
The environmental implications of real-time rendering are also noteworthy. The ability to visualize designs in various environmental contexts contributes to more sustainable architecture. Architects can assess how natural light interacts with interior spaces, optimizing energy efficiency and reducing the need for artificial lighting during the day.
In conclusion, real-time rendering has ushered in a new era of architectural design, propelling the industry into a realm of dynamic visualization, immersive experiences, and enhanced collaboration. The ability to witness designs in motion, explore different lighting conditions, and interact with virtual environments has redefined how architects approach their craft. From facilitating client presentations to fostering sustainable design solutions, real-time rendering's impact on architecture is profound and multifaceted. As the technology continues to evolve, architects have an unprecedented opportunity to push the boundaries of creativity, efficiency, and sustainability in the built environment.