Details
Joined devRant on 4/22/2017
Join devRant
Do all the things like
++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatar
Sign Up
Pipeless API
From the creators of devRant, Pipeless lets you power real-time personalized recommendations and activity feeds using a simple API
Learn More
-
I discussed using page-rank for ML a while back here - https://devrant.com/rants/11237909/...
I talk about something vaguely similar in "scoring the matches" here though - https://pastebin.com/YwjCMvRp
Incidentally the machine learning community finally caught up and did something similar on a RAG
https://news.ycombinator.com/item/... -
If you're using random in python, and need arbitrary precision, use mpmath.
If you're using it with the decimal module, it doesn't automatically convert just so you know.
Instead convert the output of arb_uniform to a string before passing to the decimal module.3 -
Turns out you can treat a a function mapping parameters to outputs as a product that acts as a *scaling* of continuous inputs to outputs, and that this sits somewhere between neural nets and regression trees.
Well thats what I did, and the MAE (or error) of this works out to about ~0.5%, half a percentage point. Did training and a little validation, but the training set is only 2.5k samples, so it may just be overfitting.
The idea is you have X, y, and z.
z is your parameters. And for every row in y, you have an entry in z. You then try to find a set of z such that the product, multiplied by the value of yi, yields the corresponding value at Xi.
Naturally I gave it the ridiculous name of a 'zcombiner'.
Well, fucking turns out, this beautiful bastard of a paper just dropped in my lap, and its been around since 2020:
https://mimuw.edu.pl/~bojan/papers/...
which does the exact god damn thing.
I mean they did't realize it applies to ML, but its the same fucking math I did.
z is the monoid that finds some identity that creates an isomorphism between all the elements of all the rows of y, and all the elements of all the indexes of X.
And I just got to say it feels good. -
I FINALLY comprehend list comprehensions.
I can write an unlimited amount of nested loops on a single line and make other less experienced people hate me for fun and profit.
Also learned about map() #I hate it#, zip(which is awesome), and the utility of lambdas (they're okay).
Enumerate is pretty nifty too, only thing I lose is setting the initial value of the iterator index.15 -
Wherein I disprove Goldbachs Conjecture (in one specific case)
golbach conjecture:
every even number is the sum of two primes
lets call the primes p and q
lets call our even number p+q=n
we can go further by establishing two additional variables
u=p-1, v=q-1
therefore every even number is the sum of u+v+2, according to goldbach's own reasoning.
in the simplest case...
p=2, q=2, p+q=4
u=1, v=1, u+v+2 = 4
We can therefore make a further conjecture in the simplest case every sum of two primes, less 2, is the sum of two composites. This likely has connections to the abc conjecture for a variety of reasons. But leaving ancillary discussion aside for a moment...
We can generalize to a statement that every even number is the sum of two odd numbers. And every odd number greater than 1 is the sum of an odd number and an even number.
Finding an even number that is not the sum of (p-1)+(q-1) would therefore be equivalent to disproving the goldbach conjecture. Likewise proving every even number is the sum of (p-1)+(q-1) would be the equivalent of proving it.
Proving all even numbers greater than 2 are the sum of two composites + 2 would be proof of goldbachs conjecture, and finding any example or an equation that proves an example exists such that *some* subset of even numbers are NOT the sum of two composites +2, would disprove the conjecture.
Lets start with a simple example:
2+2=4
because 4-2=2, and two is not the sum of two composite numbers goldbachs conjecture must ipso facto be false.
QED
If I've wildly misapprehended the math, please, somebody who is better at it, correct me.
Honestly if this is actually anything, I'd be floored to discover no one has stumbled on this line of reasoning before.8 -
Apparently inverse sigmoid is how logits are calculated.
Here I am reinventing the fucking wheel.17 -
https://simulator.io/board
Lets you place clocks, full and half adders, D latches, RS and JK flip-flops, shift registers, demultiplexers, multiplexers, and decoders, as well as all the standard gates. It also has buttons, switches, and individual LEDs.
Pretty close to what I would make myself.6 -
So I realized if done correctly, an autoencoder is really just a bootleg token dictionary.
If we take some input, and pass it through a custom hashfunction that strictly produces hashes with only digits as output, then we can train a network, store the weights and biases, and then train a decoder on top of that.
Using random drop out on the input-output pairs, we can do distillation of the weights and biases to find subgraphs that further condense this embedding.
Why have a token dictionary at all?10 -
I was never really interested in score systems until github streak. I wanted to have a streak of a year but got stuck at 40. I like this score system because it represents your productivity in someway and in top it means that you have backupped your source. It's your backup meter.
Anyone else caring about this? What's your longest streak?9 -
Unironically has "Blockchain Engineer" in their LinkedIn title...
yeah, good luck with life you fucking asshat
THAT'S RIGHT YOU ABSOLUTE SHIT BAD CODERS WHO DON'T KNOW SHIT ABOUT SHIT!!!!!!!
FULLSTACKCIRCUS IS FUCKING BACK!!!!!!!
🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡🤡
just kidding, see you again in 5 years
i'd rather keep my head down and continuing earning 5K+ passive than deal with you groveling clowns who think you have it made because you have a youtube channel - every gen Zer has like 10 of those
read >>>> sidtheitclown33 -
Tonight I've dreamed that tsoding was writing a sudoku solver. Even in my dreams I'm a nerd. Why no streamii ones or a kiki-like one? This is why I can't design anything, huge lack of imagination. I'm borderplain. It's a mental issue now. Need help and medicine5
-
Had an individual financial advisor worth 7-8 figures and with hundreds of thousands of followers, spontaneously follow me on twitter and start a conversation. He only follows a few hundred others.
Is this what it is like to meet a celebrity?
What does this person even want?
I don't know whether to be annoyed or flattered.
We're in completely different financial classes, and have nothing in common other than being trapped like rats in a cage by our own circumstances.13 -
by simply making the bias random on the second input for a two bit binary input during activation calculation, it's possible to train a neural net to calculate the XOR function in one layer.
I know for a fact. I just did it.16 -
Fuck, wanted a year long streak on github but failed at 40 days.
I did code but just had no part finished so didn't commit.
I even have an alarm for committing ffs. I just snoozed it and was like I do it when I finished x. Forgot track of time like always with coding and found out four hours later.
Fuck. Back to one. I just start over, it shouldn't be that hard.
I should change my commit behavior to push if compiles instead of push after complete feature. So, I change the definition of achievement easier to achieve6 -
Let me arrogantly brag for a moment, and let us never forget
that I front-ran GPT's o1 development by more than a week, posted
here:
https://devrant.com/rants/11257717/...
And I know what their next big development will be too. I just haven't shared it yet because it blows backpropagation out of the fucking water.
I may not be super competent at anything but I'm a god damn autistic accidental oracle when it comes to knowing what comes next in the industry.
relevant youtube video and screenshot:
https://youtu.be/6xlPJiNpCVw/...9 -
This shit is fascinating, especially reading about the variations in function of the various brodmann areas:
https://en.wikipedia.org/wiki/...
My favorite schizo-interpretation of this is "your head is full of bees."
i.e. brodmann area 10, which is thought to be responsible for memory recall strategies (basically an adaptive memory allocator and heap) has about 250 million neurons in a pinprick of a volume.
A bee has about 1 million neurons.
In otherwords: the part of your brain that decides how memory is managed has only the equivalent brainpower of 250 bees, lol.
Obviously a simplification-to-the-level-of-absurdity but it's fun to intentionally interpret something to the level of distortion.11 -
My own little version of moore's law:
In 1986 the connectome (the brain) of c. elegans, a small worm, was mapped. It would take decades before the research caught up to the point where we had the hardware to simulate it.
In 2024, we have successfully mapped, and fully simulated (to matching observed behavioral data) the brain of a fruit fly, a total of 139,255 neurons and corresponding connections.
Thats a 38 year period.
If the period is roughly 40 years, and the leap in successful neurons mapped *and simulated* is by an average of 461 times the prior number of neurons, then by 2062-2064 we will be simulating box jellyfish, fruit flys, zebrafish, bees, ants, honey bees, cockroachs, coconut crabs, geckos, guppys, sand lizards, snakes, skinks, toirtoises, frogs, iguanas, shrews, bats, and even moles.
By the dozens or hundreds in any given simulation.
By the year 2100-2104 we'll be fully simulating the brains of mice, quill, crocodiles, birds such as doves, rats, zebra finchs,
guinea pigs, lemurs, ducks, ferrets, cockatiels, squirrels, mongoose, prairie dogs, rabbits, octopi, house cats, buzzards, parakeets, grey parrots, snowy owls, racoons, and even domestic pigs.
And in the years between 2100 to 2140, starting immediately with domestic dogs, we will ramp up and end with the capacity to simulate human brains in full, probably by the dozens or hundreds.
This assumes we can break the quantum barrier of course.20 -
This morning I woke up because some light from the hallway was comming from under my door. I went on investigation. Was a bit scared. But it was just Kiki sitting there solving a rubiks cube while speaking UTF-8 to herself. I went back to bed6
-
Deleted over 1'500 lines of code over the last 2 days and replaced it with 80 lines of readable, simple, generic code.
And I'm feeelin' gooood 🎶7 -
Working two hours on this FFS. Three the same laptops:
- Two USB sticks of different brands working on two of the laptops. One doesn't work.
- BIOS versions: the working two are from 2018 and 2020. The not working one is from 2022.
- BIOS settings: 99% the same, especially where matters. Literally went trough every menu.
- I thought, maybe the 'new' 2022 BIOS has a buggy - so maybe update BIOS? Everything only for windows on Lenovo website.
I installed xubuntu on it before. All laptops say "cant find /boot" but on two of them it's not a problem and they run the live USB stick with option to installii. Since I installed it before, the BIOS version is probably not the issue.
If i close my eyes i see swastika's.
Detail: the not working laptop is the one that i wrote the xubuntu iso to /dev/sda (what was the hard drive, see a few rants ago). For some reason, it aggresively boots from that one. I do see my USB stick working (very busy flashing light). Is it maybe possible that it mounts my HD as installation cdrom? The HD contains those files.
Anyone tips?2 -
Wherein I bait reddit into over 70k views with some bullshit about AI:
https://reddit.com/r/Futurology/...
I almost wonder if their viewcount/usercount is real.
The model is real, my consultant background is not.5 -
A kind of verbose discussion of my earliest ideas and discussion with Nous LLM (Claude) about my new NAS/CL LLM model:
https://pastebin.com/YwjCMvRp2 -
You know the old adage "cost, speed, quality, pick two."
I've come up with my own.
Shitty jobs you'll forever be stuck in with no way out: unreasonable demands and coworkers that drive you insane, pay below the poverty line, no sleep.
Pick two.
More likely, pick three.6 -
Some notes from prior to developing my current language model:
https://miro.com/app/board/...
Started with ngrams, moved on from that, and the whole thing got away from me fast.
Working on building and training it on rgb-to-color categorization this week. Experiments designed just gotta implement it now.1 -
Programming smth complex is literally time traveling. My clock goes in these steps:
20:00 -> so much time left
22:00 -> time to wrap up soon
02:00 -> wut? Impossible
The type of projects I do is quite limited to standard library so I also don't spend time on reading docs of some dependency. I make everything myself. Afterwards, I look up what the way to go is how others do it and compare. Mostly, I don't adjust, it's just smth for next time. That's the whole hobby. Just keep going6 -
Remember my LLM post about 'ephemeral' tokens that aren't visible but change how tokens are generated?
Now GPT has them in the form of 'hidden reasoning' tokens:
https://simonwillison.net/2024/Sep/...
Something I came up with a year prior and put in my new black book, and they just got to the idea a week after I posted it publicly.
Just wanted to brag a bit. Someone at OpenAI has the same general vision I do.15