Join devRant
Do all the things like
++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatar
Sign Up
Pipeless API
From the creators of devRant, Pipeless lets you power real-time personalized recommendations and activity feeds using a simple API
Learn More
Search - "phenomenon"
-
I as the "coffeedrinker" stopped drinking coffee and energy drinks. I do not want to be anyone's boss but if you think you can live without them, do it. I am fairly young and after a long period of frequent consumption ( 1 coffee or drink per day), I was waking up at nights or struggling during the days from constant heart pains. Now 2 months later this phenomenon happens seldom and I feel better and more refreshed after my sleeping. I know this is irrelevant but I know fellow devs that overconsume these kind of drinks. At least if you can limit your dose! :) I just want everyone to be healthy and happy! Have a nice day! ♥58
-
Should I Close-Source my project?
I have been working on a Desktop/hacking simulator game and up until now the project has been Open-Source. I'm at a point now where I haven't gone too far to turn back.
Last night I got to thinking about my game, and what I want to do in the future. The game will always remain Free, but I might sell it to another company later down the line, something I can't do if I stay Open. I want to makea good game. And I don't want to do it for money (because that has never worked out for me in the past) but I want to *be able* to make money if I wanted to. I mean, I have been told by several developers that my game will be "ground breaking/a worldwide phenomenon/a Minecraft competitor" while being Open is one of my main selling points, besides populatity, what do I have to gain? I said I don't want to develop for money (mainly because the pressure gets to me) but I'm so poor I'm almost literally starving. I make $3/mo from Patreon and survive from donation from relatives. I feel like I need this. But I also feel selfish. Information should be free, ya know?
Idk.. This started serious and turned into a ramble.. Guess that's what this app is all about.
Leave your opinions below.25 -
Haha kids, you're all dead wrong. Here's my story.
There is a thing called “emergence”. This is a fundamental property of our universe. It works 100% of the time. It can't be stopped, it can't be mitigated. Everything you see around you is an emergent phenomenon.
Emergence is triggered when a lot of similar things come together and interact. One water molecule cannot be dry or wet, but if you have many, after a certain number the new property emerges — wetness. The system becomes _wet_.
Professionalism is an emergent phenomenon too, and its water molecules are abstract knowledge. Learn tech things you're interested in, complete random tutorials, code, and after a certain amount of knowledge molecules is gained, something clicks inside your head, and you become a professional.
Unfortunately, there are no shortcuts here. Uni education can make you a professional seemingly quicker, but it's not because uni knowledge is special, it's because uni is a perfect environment to absorb a lot of knowledge in a short period of time.
It happened to me too. I started coding in Pascal in fifth grade of high school, and I did it till sixth. Then, seventh to ninth were spent on my uni's after-school program. After ninth grade, I drop out of high school to get to this uni's experimental program. First grade of uni, and we're making a CPU. Second grade, and we're doing hard math, C and assembly.
And finally, in the third grade, it happens. I was sitting there in the classroom, it was late, and I was writing a recursive sudoku solver in Python. And I _felt_ the click. You cannot mistake it for anything else. It clicks, and you're a changed person. Immediately, I realized I can write everything. Needless to say, I was passing everything related to code afterwards with flying colours.
From that point, everything I did was just gaining more and more experience. Nothing changed fundamentally.
Emergence is forever. If you learn constantly, even without a concrete defined path, I can guarantee you that you _will_ become a professional. This is backed by the universe itself. You cannot avoid becoming one if you're actively accumulating emergence points.
Here's the list of projects I made in the past 11 years: https://notion.so/uyouthe/...
I'm 24.7 -
Psychic readings https://linkedin.com/pulse/... are one of the most mysterious and fascinating areas of the paranormal. This phenomenon has long attracted the attention of both ordinary people and scientists, since it represents the ability to receive information in unusual ways, bypassing the usual five senses.
Psychics, or people with such abilities, claim that they can sense energetic interactions, see objects and events at a distance, read thoughts, obtain information about a person only from his photograph, and so on. One of the most well-known psychic readings is tarot card reading, which allows psychics to predict the future and give advice on decision-making.
There are many theories about how psychic readings work. Some believe that psychics are able to perceive information not only through the usual five senses, but also through the sixth sense - intuition. Others believe that psychic abilities are related to a person's energy fields and aura.
In order to understand this phenomenon, scientists conduct numerous studies and experiments. However, it has not yet been possible to find a scientific explanation for extrasensory abilities. Some experiments show that psychics can detect information that ordinary people cannot see, but this has not yet been scientifically proven.
Many people turn to psychics in search of answers to questions regarding their personal life, career, health and other important aspects. Psychics offer them consultations and help them understand difficult situations, predict the future and help them make important decisions.
However, it is worth remembering that there are many impostors and scammers who try to use the popularity of psychic abilities to deceive. Therefore, it is important to choose trusted specialists and not get hung up on the predictions and advice of psychics, but make decisions independently, based on your own judgment and intuition.
Overall, psychic readings remain a mystery to science and society. Many people are confident in the reality of such abilities, others consider them fiction and deception. However, whether you believe in psychic abilities or not, it is worth recognizing that these paranormal phenomena continue to attract the attention and interest of many people around the world.6 -
When you realise that you no longer interpret s3 as a smartphone model, fork as an eating tool, eclipse as natural phenomenon... :( :|2
-
A rare weather phenomenon called "snezhura" (slush) was observed in St. Petersburg recently. I don't quite get how it works, yet it's still fascinating.10
-
DEVFALL -
phenomenon of hair fall in the budding developer.
Cause :- Staying awake 48 hours in a day
Cure :- CURE? WHY'D YOU WANT A CURE, YOU ARE ABOUT TO BECOME THE ULTIMATE BEING.
But seriously, cure :- just fucking sleep 9 hours a day for 6 months2 -
To be honest, I'm not as excited as I was 6-7 years ago when our tech industry seen a big leap, where these ML/Deep Learning algorithms were out performing humans, Apache Spark out perfomed Hadoop in distributed computing, Docker/Kubernetes are the new phenomenon in software development and delivery, Microservices architecture, ReactJS virtual DOM concepts were so cool.
Really though, I've come realise that these software trends come and go. All you need to do is adapt and go with the flow.3 -
The time that we dedicate to the things and people that we love/like, when it's enough?
The question is generic and for good reason.
Yesterday, semi-seriously, my gf asked me when we'll have a baby, I answered, seriously, that it's gonna be when I'll feel ready to share the daily time with someone as demanding as another family member growing up.
Now, between job time, hobbies time and girlfriend (gonna marry soon) time the time is already tight and because I'm self sufficient about happiness and kind of a loner I don't share really much time with her most of the days, and from this realisation from her side she broke into crying.
From that experience I understood that there might be need some adjustment on my side.
But on another side I'm puzzled of how other families deal with this, because though my life I've seen couples/married-people that had not really much interactions with each other on a daily basis and seemed fine with living like that.
So knowing this context, what's your experience about this phenomenon through your life time?4 -
Am I the only one that can't stand non-maximized windows on Windows but hate maximized windows on Mac?
Could this be a psychological phenomenon in which the designs of these systems make you use them in a specific way? Or am I just retarded?
Anyone else does the same?8 -
Without Unix, there would have been no Minix (Tanenbaum et al.) orGNU (Richard Stallman et al.).Without Minix, there would be no inspiration to write Linux. Remember that Linus started his “project” because he didn’t like many of the design decisions Tanenbaum has taken in Minix, including the microkernel. In fact, Linus has tried to submit some changes to the professor and the latter rejected them. So the young chap decided to write his own kernel using his design.Without GNU, there would be no open source tools that Linus himself used to write, compile, test and distribute his project, to become a few years later a global phenomenon. Also, the fact that GNU was already an established Unix clone (minus an operating kernel) at that time helped Linus to focus on the missing part, the kernel. Otherwise, he would not have known where to start.And finally, Unix was the template all of the above (and more) were trying to imitate. Without it, there would have been nothing to clone from.1
-
Sometimes i cant fix a bug for days. Like 3 days of brainwreck. Then on the 4th day i wake up. So whatever i love. Take some time for a rest. And then begin working whenever i feel like it. I start working at 2 pm. Try to solve the same bug again. The first thing that comes to my mind is Hold on, why dont i try to change this? I did and it worked. My first thought has solved a 3 day old bug.
Can someone explain this phenomenon. This is proof that a man is unproductive and cant work good if he doesnt feel like it.
You know all of those bullshit andrew tate quotes "i work even when I don't feel like it because that's what men are supposed to do. I train when im happy and i train exactly the same when im unhappy" but thats bullshit. I can not be productive if i am unhappy. I tried so hard and the harder i tried the more i failed. And now when im no longer unhappy i solved it on the first try.
Nobody cares when a man is unhappy. No one gives a shit. It's not fair1 -
Is there a name for the phenomenon whereby you iteratively modify code to try to fix a bug, with no apparent result, and then realize it's an entirely different part of the code causing the issue, but the parts you were modifying actually did need modifying too?2
-
Thinking of (the possible myth?) that phenomenon where you can ingest small doses of poison to build up an immunity over time, I'm convinced energy drinks are released by the government to build up our immunity to toxic bullshit because holy fuck I have never felt good during or after drinking one of those fucking things.5
-
Shouldn't there be a phenomenon named "critically timed bug" which always shows up like '15 mins before sleep time', '10 mins before your workplace's leaving time' or even worse '5 minutes before production/presentation'?
-
Curious question to fellow Mint Mate user (if there is any).
Do you experience a long restarting phenomenon with your mint?
I have used this and that distros and windows in the past. I am fond of light and fast things; long time user of xubuntu and lubuntu. Mint saved my ass when I had some display driver issues with my xubuntu. So I have switched to and I have been using it for a while and noticed that sometimes the restarting process can take more than 90 second.5 -
Sydochen has posted a rant where he is nt really sure why people hate Java, and I decided to publicly post my explanation of this phenomenon, please, from my point of view.
So there is this quite large domain, on which one or two academical studies are built, such as business informatics and applied system engineering which I find extremely interesting and fun, that is called, ironically, SAD. And then there are videos on youtube, by programmers who just can't settle the fuck down. Those videos I am talking about are rants about OOP in general, which, as we all know, is a huge part of studies in the aforementioned domain. What these people are even talking about?
Absolutely obvious, there is no sense in making a software in a linear pattern. Since Bikelsoft has conveniently patched consumers up with GUI based software, the core concept of which is EDP (event driven programming or alternatively, at least OS events queue-ing), the completely functional, linear approach in such environment does not make much sense in terms of the maintainability of the software. Uhm, raise your hand if you ever tried to linearly build a complex GUI system in a single function call on GTK, which does allow you to disregard any responsibility separation pattern of SAD, such as long loved MVC...
Additionally, OOP is mandatory in business because it does allow us to mount abstraction levels and encapsulate actual dataflow behind them, which, of course, lowers the costs of the development.
What happy programmers are talking about usually is the complexity of the task of doing the OOP right in the sense of an overflow of straight composition classes (that do nothing but forward data from lower to upper abstraction levels and vice versa) and the situation of responsibility chain break (this is when a class from lower level directly!! notifies a class of a higher level about something ignoring the fact that there is a chain of other classes between them). And that's it. These guys also do vouch for functional programming, and it's a completely different argument, and there is no reason not to do it in algorithmical, implementational part of the project, of course, but yeah...
So where does Java kick in you think?
Well, guess what language popularized programming in general and OOP in particular. Java is doing a lot of things in a modern way. Of course, if it's 1995 outside *lenny face*. Yeah, fuck AOT, fuck memory management responsibility, all to the maximum towards solving the real applicative tasks.
Have you ever tried to learn to apply Text Watchers in Android with Java? Then you know about inline overloading and inline abstract class implementation. This is not right. This reduces readability and reusability.
Have you ever used Volley on Android? Newbies to Android programming surely should have. Quite verbose boilerplate in google docs, huh?
Have you seen intents? The Android API is, little said, messy with all the support libs and Context class ancestors. Remember how many times the language has helped you to properly orient in all of this hierarchy, when overloading method declaration requires you to use 2 lines instead of 1. Too verbose, too hesitant, distracting - that's what the lang and the api is. Fucking toString() is hilarious. Reference comparison is unintuitive. Obviously poor practices are not banned. Ancient tools. Import hell. Slow evolution.
C# has ripped Java off like an utter cunt, yet it's a piece of cake to maintain a solid patternization and structure, and keep your code clean and readable. Yet, Cs6 already was okay featuring optionally nullable fields and safe optional dereferencing, while we get finally get lambda expressions in J8, in 20-fucking-14.
Java did good back then, but when we joke about dumb indian developers, they are coding it in Java. So yeah.
To sum up, it's easy to make code unreadable with Java, and Java is a tool with which developers usually disregard the patterns of SAD. -
Apart from the fact that I arrived at a good framework at work to play in problem space than in solution space, this post is more about self realisation and a slight progress in my happiness levels.
Monsoons started in India. The vibe somehow had always been melancholic for me triggering SAD (aka seasonal depression).
However, this year I find it cosier than ever. Hot showers, lazing around on a holiday when it's pouring outside, watching my favourite show/movie. I feel very relaxed in the moment, even when work and life is not as expected/under control.
What I realised is that my problem can be solved. I need a bigger house. That would give me privacy, some personal space for hobbies, and put a barrier between me and parents easing the tension and clashes. I could then get married, and with all the money I will save (from not buying a house myself), can be used to pursue hobbies like music, art, travel, etc.
Whenever I relax, my sleep pattern changes where I have longer duration of deep sleep with many dreams (perhaps processing everything). Does anyone else experience such a phenomenon?
Anyway, life doesn't get easy or hard, we just learn to put up with shit.4 -
52.7% software engineers experience frequent-intense levels of imposter phenomenon. Women (60.6%) >> men (48.8%). Asian (67.9%) and Black (65.1%) >> White (50.0%). Less common if married with kids. Guenes et al. at ICSE 2024.
Preprint https://arxiv.org/pdf/...15 -
Warning - Not IT related.
Long ago, I had a chemist professor who told us a story. She worked in a laboratory where they have studied cristal formations, so basically made a liquid highly capable to form cristals, and they watched them forming, doing tests and so on. In the meantime new building of the campus opened and they had to move the lab to the new location, which was a fourth floor of newly made building. Few of them started to work there even before they moved the old materials and equipment and they started few cristalisation studies, the interesting part is that the cristals didn't formed. She said that at the end they had many cups with prepared liquid and apsolutely no cristals for weeks, but one day the lead researcher arrived with the old, already formed cristals, from the old lab, and toon those inside of the room with prepared cups all the cristals started to form at the same time. After telling us a story she asked us not to tell this to anyone because the science currently doesn't accept this phenomenon and we will be demonised and looked fools it the scientific community.
This story made a hole in my brain...
It was like 10 years ago, and as a problem solver I still have sometimes some weird ideas about it, and strange explanations comming from nothing, and without any deep understanding of quantum physics or even cristalisation. :D1 -
Ever been in that scenario where your mind, with all its white and grey cells in you decide as a singular unit to do something.,
Mind: Let's do something productive for the weekend
Me: Like what?
Mind: Lets program a Global phenomenon in two days.
Me: ok. If you say so.
(Opens browser)
(Notification - <printsomenounhere>.js is released.)
Mind: FUCK
Me: YAY (Endless Wormhole printsomenounhere.js -> printanothernoun.js -> you know the drill)2 -
Russians Engineer a Brilliant Slot Machine Cheat
...But as the “pseudo” in the name suggests, the numbers aren’t truly random. Because human beings create them using coded instructions, PRNGs can’t help but be a bit deterministic. (A true random number generator must be rooted in a phenomenon that is not manmade, such as radioactive decay.) PRNGs take an initial number, known as a seed, and then mash it together with various hidden and shifting inputs—the time from a machine’s internal clock, for example—in order to produce a result that appears impossible to forecast. But if hackers can identify the various ingredients in that mathematical stew, they can potentially predict a PRNG’s output. That process of reverse engineering becomes much easier, of course, when a hacker has physical access to a slot machine’s innards...
https://wired.com/2017/02/...1 -
Enlightenment did happen, but it would be a reach to call a little rich Italian hipsters’ literature fad a “worldwide phenomenon”.
When we emerged as a species, all we had was tools and fear. Nothing has changed. Our progress in tools did fix suffering somewhat, but it couldn’t fix fear.
Fear is what makes one person attack the other, from a pub scuffle to launching nukes. It’s all the same, isn’t it. If I don’t strike now, they’ll strike first.
Losing an argument says nothing about you. Someone yelling at you while you stay silent says nothing about you. Being rejected says nothing about you. Being ghosted says nothing about you. Being betrayed says nothing about you. Even obeying your boss says nothing about you.
There is no need to compensate. You have the power to turn your “yes” into “no” swiftly and confidently whenever you want to. -
The Turing Test, a concept introduced by Alan Turing in 1950, has been a foundation concept for evaluating a machine's ability to exhibit human-like intelligence. But as we edge closer to the singularity—the point where artificial intelligence surpasses human intelligence—a new, perhaps unsettling question comes to the fore: Are we humans ready for the Turing Test's inverse? Unlike Turing's original proposition where machines strive to become indistinguishable from humans, the Inverse Turing Test ponders whether the complex, multi-dimensional realities generated by AI can be rendered palatable or even comprehensible to human cognition. This discourse goes beyond mere philosophical debate; it directly impacts the future trajectory of human-machine symbiosis.
Artificial intelligence has been advancing at an exponential pace, far outstripping Moore's Law. From Generative Adversarial Networks (GANs) that create life-like images to quantum computing that solve problems unfathomable to classical computers, the AI universe is a sprawling expanse of complexity. What's more compelling is that these machine-constructed worlds aren't confined to academic circles. They permeate every facet of our lives—be it medicine, finance, or even social dynamics. And so, an existential conundrum arises: Will there come a point where these AI-created outputs become so labyrinthine that they are beyond the cognitive reach of the average human?
The Human-AI Cognitive Disconnection
As we look closer into the interplay between humans and AI-created realities, the phenomenon of cognitive disconnection becomes increasingly salient, perhaps even a bit uncomfortable. This disconnection is not confined to esoteric, high-level computational processes; it's pervasive in our everyday life. Take, for instance, the experience of driving a car. Most people can operate a vehicle without understanding the intricacies of its internal combustion engine, transmission mechanics, or even its embedded software. Similarly, when boarding an airplane, passengers trust that they'll arrive at their destination safely, yet most have little to no understanding of aerodynamics, jet propulsion, or air traffic control systems. In both scenarios, individuals navigate a reality facilitated by complex systems they don't fully understand. Simply put, we just enjoy the ride.
However, this is emblematic of a larger issue—the uncritical trust we place in machines and algorithms, often without understanding the implications or mechanics. Imagine if, in the future, these systems become exponentially more complex, driven by AI algorithms that even experts struggle to comprehend. Where does that leave the average individual? In such a future, not only are we passengers in cars or planes, but we also become passengers in a reality steered by artificial intelligence—a reality we may neither fully grasp nor control. This raises serious questions about agency, autonomy, and oversight, especially as AI technologies continue to weave themselves into the fabric of our existence.
The Illusion of Reality
To adequately explore the intricate issue of human-AI cognitive disconnection, let's journey through the corridors of metaphysics and epistemology, where the concept of reality itself is under scrutiny. Humans have always been limited by their biological faculties—our senses can only perceive a sliver of the electromagnetic spectrum, our ears can hear only a fraction of the vibrations in the air, and our cognitive powers are constrained by the limitations of our neural architecture. In this context, what we term "reality" is in essence a constructed narrative, meticulously assembled by our senses and brain as a way to make sense of the world around us. Philosophers have argued that our perception of reality is akin to a "user interface," evolved to guide us through the complexities of the world, rather than to reveal its ultimate nature. But now, we find ourselves in a new (contrived) techno-reality.
Artificial intelligence brings forth the potential for a new layer of reality, one that is stitched together not by biological neurons but by algorithms and silicon chips. As AI starts to create complex simulations, predictive models, or even whole virtual worlds, one has to ask: Are these AI-constructed realities an extension of the "grand illusion" that we're already living in? Or do they represent a departure, an entirely new plane of existence that demands its own set of sensory and cognitive tools for comprehension? The metaphorical veil between humans and the universe has historically been made of biological fabric, so to speak.7