Join devRant
Do all the things like
++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatar
Sign Up
Pipeless API
From the creators of devRant, Pipeless lets you power real-time personalized recommendations and activity feeds using a simple API
Learn More
Search - "ai funding"
-
The best way to get funding from VCs now is to include the following words: ML, AI, IoT. To even blow their minds more, add Blockchain.2
-
Most people will know how the AI in 2001 A Space Odyssey came to be called HAL. By the simple expedient of taking the letters IBM and applying -- to each.
How many remember how the Windows have all come to know and love started out as Windows NT?
It came about when Digital pulled the funding for the new version of it's VMS operating system, and Bill Gates swooped in and hired Dave Cutler, who basically took all the code with him to make the new version of Windows.
And stick the finger to his former employer, incremented VMS to get WNT. -
This is gonna be a long post, and inevitably DR will mutilate my line breaks, so bear with me.
Also I cut out a bunch because the length was overlimit, so I'll post the second half later.
I'm annoyed because it appears the current stablediffusion trend has thrown the baby out with the bath water. I'll explain that in a moment.
As you all know I like to make extraordinary claims with little proof, sometimes
for shits and giggles, and sometimes because I'm just delusional apparently.
One of my legit 'claims to fame' is, on the theoretical level, I predicted
most of the developments in AI over the last 10+ years, down to key insights.
I've never had the math background for it, but I understood the ideas I
was working with at a conceptual level. Part of this flowed from powering
through literal (god I hate that word) hundreds of research papers a year, because I'm an obsessive like that. And I had to power through them, because
a lot of the technical low-level details were beyond my reach, but architecturally
I started to see a lot of patterns, and begin to grasp the general thrust
of where research and development *needed* to go.
In any case, I'm looking at stablediffusion and what occurs to me is that we've almost entirely thrown out GANs. As some or most of you may know, a GAN is
where networks compete, one to generate outputs that look real, another
to discern which is real, and by the process of competition, improve the ability
to generate a convincing fake, and to discern one. Imagine a self-sharpening knife and you get the idea.
Well, when we went to the diffusion method, upscaling noise (essentially a form of controlled pareidolia using autoencoders over seq2seq models) we threw out
GANs.
We also threw out online learning. The models only grow on the backend.
This doesn't help anyone but those corporations that have massive funding
to create and train models. They get to decide how the models 'think', what their
biases are, and what topics or subjects they cover. This is no good long run,
but thats more of an ideological argument. Thats not the real problem.
The problem is they've once again gimped the research, chosen a suboptimal
trap for the direction of development.
What interested me early on in the lottery ticket theory was the implications.
The lottery ticket theory says that, part of the reason *some* RANDOM initializations of a network train/predict better than others, is essentially
down to a small pool of subgraphs that happened, by pure luck, to chance on
initialization that just so happened to be the right 'lottery numbers' as it were, for training quickly.
The first implication of this, is that the bigger a network therefore, the greater the chance of these lucky subgraphs occurring. Whether the density grows
faster than the density of the 'unlucky' or average subgraphs, is another matter.
From this though, they realized what they could do was search out these subgraphs, and prune many of the worst or average performing neighbor graphs, without meaningful loss in model performance. Essentially they could *shrink down* things like chatGPT and BERT.
The second implication was more sublte and overlooked, and still is.
The existence of lucky subnetworks might suggest nothing additional--In which case the implication is that *any* subnet could *technically*, by transfer learning, be 'lucky' and train fast or be particularly good for some unknown task.
INSTEAD however, what has happened is we haven't really seen that. What this means is actually pretty startling. It has two possible implications, either of which will have significant outcomes on the research sooner or later:
1. there is an 'island' of network size, beyond what we've currently achieved,
where networks that are currently state of the3 art at some things, rapidly converge to state-of-the-art *generalists* in nearly *all* task, regardless of input. What this would look like at first, is a gradual drop off in gains of the current approach, characterized as a potential new "ai winter", or a "limit to the current approach", which wouldn't actually be the limit, but a saddle point in its utility across domains and its intelligence (for some measure and definition of 'intelligence').4 -
[Background]
Back in September I joined a startup after my first job in MNC for about 1.8 yrs as a fresher. I always wanted to learn, but the experience in that MNC was not at all fruitful. So ai decided to join a small/mid size company or a startup. To my luck, I got in this small startup in a week after my resignation as a front-end dev (always wanted to be).
It's an automation company, so you can find software, electronics, even mechanical engineer.
The team was almost a year younger than me. It was a team of around 12 people, in which 5 of them were from Business development.
The tech team was too driven and knowledgeable. Always trying new stuffs and motivating to do the same. I was highly motivated by them in my initial days, watching them working on new stuffs.
So I started with revamping their website completely in Angular 4, and did it in around a month or so, being new to Angular. Outcome was pretty satisfactory. I wanted to work on new projects, but just to get the cashflow in they started getting in WordPress projects. It was frustrating, I wanted to work more on new technologies like Angular, React, etc...but just for the survival of the company I had to work on WordPress, so to respect their urge to get going I kept working on 3-4 projects in parallel, and mind you the clients were from hell !!
Fast-forward 4 months, I am still working on few WordPress websites, and one internal GPS based project in React. And I haven't received my salary for past 3.5 months, since the company is still struggling with the issue of funding and getting money from clients. I kinda liked working there because there was lot to learn even though they are so young, but I had bills to pay too.
And I am in dilemma to leave the company or not, because I already stretched 3 months out of good will and guilt of leaving the company in high time. So i finally let the CEO know that I cannot stick for any longer. And i was done with the false promises of getting the salary "next month" everytime. All the money getting inside of company was invested heavily on the product we were building and no one was getting the salaries. Others were fine since they were founding members too.
Long story short : I finally left immediately and now working in a good company as a React dev. I hope they do well and I would love to see them grow, but please *STOP* making false promises and hold on to employees on a lie.1 -
"AI could be used to make chemical, biological, nuclear weapons" they keep saying ad nauseam
it isn't like those things are difficult to make
it's like anthropic wants to advertise for people to use their AI to sow chaos in the world. they are doing a deal with Amazon which had CIA funding historically
there is a depopulation cult and they ain't stopping
"nooo, don't go into my bedroom senpai" says the porno. gosh wonder what the advertisement is for6