Do all the things like ++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatarSign Up
From the creators of devRant, Pipeless lets you power real-time personalized recommendations and activity feeds using a simple APILearn More
Search - "nextcloud"
Companies can't use free software?
Well, I just forced everyone to use Nextcloud instead of Dropbox and Google Drive. People did not like it as first but the quickly realized how good it was.
We also moved from skype, teams and slack to matrix.
I am glad that matrix-synapse is so easy to install nowadays!19
Definitely Nextcloud. Now I don't have to use proprietary software like Google drive and Dropbox anymore.
It's a shame I found out about it last week. I wish I found it sooner.11
Friend's site, mail verification, nextcloud etc. all went down...
Checked all his servers, all his configs and what not... Just to realize the moron forgot to pay his bills (so his domain expired)...5
"Hmm, I wonder if Cloudflare has a limit to the amount of data I can store in the CDN"
*Uploads a shitton of crap to a Nextcloud server and caches everything via cloudflare*
So far - 300GB stored.4
I would like to invite you all to test the project that a friend and me has been working on for a few months.
We aim to offer a fair, cheap and trusty alternative to proprietary services that perform data mining and sells information about you to other companies/entities.
Our goal is that users can (if they want) remain anonymous against us - because we are not interested in knowing who you are and what you do, like or want.
We also aim to offer a unique payment system that is fair, good and guarantees your intergrity by offer the ability to pay for the previous month not for the next month, by doing that you do not have to pay for a service that you does not really like.
Please note that this is still Free Beta, and we need your valuable experience about the service and how we can improve it. We have no ETA when we will launch the full service, but with your help we can make that process faster.
With this service, we do want to offer the following for now:
Nextcloud with 50 GB storage, yes you can mount it as a drive in Linux :)
Email Client that you can connect to your email service (
Talk ( voice and video chat )
Mirror for various linux distros
We are using free software for our environment - KVM + CEPH on our own hardware in our own facility. That means that we have complete control over the hosting and combined with one of the best ISP in the world - Bahnhof - we believe that we can offer something unique and/or be a compliment to your current services if you want to have more control over your data.
Feel free to user our mirror:
Please send your feedback to:
I am looking for testers!
operationtulip.com is a small startup that is providing storage with Nextcloud. Currently every user gets 50 GB for free for as long as we are able to provide it.
We would like freeback from our users aswell:
Thank you :)30
ALMOST HALF AN HOUR SPENT TRYING TO LOG INTO MY FUCKING RASPBERRY PI OVER SSH.
you know what the problem is?
I’m not gonna tell you because I want you to feel the agony too.
> be me
> want to set up a nextcloud instance on pi to play with
> boot up
> ssh pi
^tries like 60 more times with different things
> pulls HDMI out of PC
> connect to pi direct
> check the logs
>try login from phone
>Tries resetting password
>tried logging into other pi
>type @ sign
Hey there 👋
I am more or less throwing any burden (WhatsApp, Facebook, Google etc.) out of my life. Of course I will continue using the Google account for YouTube and some games that need it.
That's what it looks like right now:
Raspberry Pi 3B+
- forum - complete (atm just for me)
- blog - no ideas and just installed october cms and nothing done yet
- nextcloud - complete and filled with my porn... eeh... data
(missing spamassassin, clam or sth. like this but it's working 😂)
(as an additional alternative to messengers)
Raspberry Pi 2
✅ catches dust
Of course, many more configurations and the like are necessary before everything is ready... but what then or what else is there?
At the moment I still use WhatsApp. Just wanna take time before sending everyone a message about changing the messenger and that it should be important for thinking about the own privacy, which alternatives there are bla...
Edit: For passwords I'm using Myki - didn't hear anything bad about it yet and it's very easy to use (Firefox add-on, Android app).
I love my passwords with 200 characters 😂
Maybe someone's knowing more about them?
Hope I didn't forget a thing... thanks in advance aaaaaaand... I'm gone. ☺23
I have multiple (in no particular order) :
Nextcloud : It was an idea that I had in my head as well - to take on corporations like Google in the space of personal cloud. Be free, open-source and put the users in charge.
Gitlab : The most open and transparent company that I've ever come across. And they work 99% remote. They've got features that no other players in the space have. All while putting users in control.
Fediverse social media - Mastodon, GNU Social, Diaspora (soon) : For taking a major step in the direction of putting the users in control of their data; all while enabling a decentralized social network.
Ruby : An open community and building a programming language that runs a lot of software of the world.
Python : The oldest thriving community that has a special place in the development community (and my heart)
The VPS where I host my owncloud instance has had its service tier upgraded since I spun it up. Opening a new one for the same price give you much better hardware stats now. Also, it's running Debian 8, which no longer gets updates from the owncloud repo. So yesterday I took this opportunity to make a new VPS, and try out nextcloud with it. I am floored. It is so much simpler for me. I'm not locked into whatever backend nextcloud chooses-- I have a choice.
Also, I can set up Apache to work however I want. I assume it'll work fine with nginx too.
Once it was installed and ready to go, I noticed I have much more granular and controlled access to my settings.
I used to love mozilla as a community. The mozilla foundation loved the community and always listened to their voice. Of recent, they've started to turn a deaf ear to the community's voices and have started moving the agenda for a corporate culture. They slowly killed many amazing projects that was mostly run by the community (RIP Firefox OS) and started focusing on more corporate-oriented lobbying and agenda (*cough* Pocket *cough*). I feel that the company is slowly moving toward becoming less of a (community-oriented) foundation and more of a corporation. That path is dangerous and not one that I expected mozilla to take.
Another company worth mentioning was OwnCloud which got forked into Nextcloud because they didn't care about the community enough and put the enterprise customers and their needs ahead of the community's. The disappointed founder of the OwnCloud quit and forked it into NextCloud with the right controls for the community and the users to always be put in the first position.19
On the one hand I really wish there was a good open-source self-hosted file sharing and collaboration service which made use of modern web technologies to provide most of its features inline as well and not just through APIs consumed by client applications.
On the other hand, I tried writing a WebDAV server once, and I fully empathize with the people who have to deal with this madness.
Nice, nextcloud just added support for ActivityPub and that means federation against Mastodon, GNUSocial, Pleroma and soon Diaspora10
i wrote a website, a server in go, a small os in c, a game in js, a game and server and web scraper and other desktop apps in java, mobile apps with flutter, a website with php also, implemented aes in go, wrote a parser in java. done sysadmin stuff on my vps and pihole/openvpn/nextcloud on my rpi. learn about c vulnerabilities and used metasploit. attempted to write an interpreted language. did some led displays with arduino. currently learning tensorflow.
i have never...
- written a driver
- made a game with a game engine
- created a file encoding
- implemented an oauth2 server
- made an api
- worked with vr
what am i missing? i want to be a very well rounded dev.15
Ubuntu, Openbox, vim, duckduckgo, Gmail, mailgun, digitalocean, xterm, libvirtd, remmina, polipo, insomnia, ulauncher, copyq, nextcloud, rofi, ssh, bash, Firefox, Firefox-Dev, Vivaldi, steam, itch, git, proton, wine, vlc, cherry mx brown and black, android, mint mobile, Asus, amd, ubiquiti, and plex.
What's in your workflow?4
WHY NEXTCLOUD, WHY DOES IT HAVE TO BE SO FUCKING COMPLICATED TO UPGRADE YOU TO THE LATEST VERSION??????
NOT ONLY DID THE DATABASE MIGRATIONS NOT RUN, YOU DIDN'T HAVE THE UPDATED VERSION OF THE FUCKING AUTHENTICATTION PROCESSOR IN THE OFFICIAL SERVER DOWNLOAD SO THE WHOLE THING WAS BROKEN AFTERWARDS.
I'VE WASTED 4 HOURS ON THIS FUCKING UPGRADE 😤😤😤😤😤😤😤4
Setting up nextcloud on a raspberry pi taught me more than 1 semester of networks/system administration course in university. This is sad.5
No actual data loss here, but the feeling of data loss.
After having my data scattered across several devices i decided to get a grip on it use a cloud. I'm too paranoid for a real cloud so i used a local nextcloud installation. That was done via docker and with a 2TB raid1-array.
I noticed that after restarting the server the cloud was somehow reset and pointed me to the setup-page, afterwards my files were already there. It did strike me as odd but i figured "maybe don't restart the server in the next time".
But i did restart it. And this time i had to setup the cloud again, but my files were gone. I got close to a heart attack, even though all those files weren't that valuable. I ripped one disk from the usb hub, connected it to my laptop and tried to mount it, but raid array. Instead i started photorec and recovered a bunch of files, even though their names were some random hex and i knew i'd spend my next weeks sorting my files. While photorec ran i inspected the docker container and saw that there were only 10GB of space available. After a while and one final df i found the culprit: the raid. For some reason the raid wasn't mounted at boot and docker created the volumes on the servers hard disk, same goes for the container data. After re-adding the disk to the hub i mounted the raid and inspected everything again. All my files were still there.
At no point did i lose my data, but the thought was shocking enough. It'd be best not to fiddle with this server in the next time.
I actually want to rant and pitch a product idea here:
So I am furious when i found out Gravit Designer made offline mode premium only, seriously what the fuck. And Figma is getting no progress for the offline mode either - even lacking grids that I need (I might do isomorphic and I also need grids to keep things aligned).
To be honest I liked Figma more now than Gravit (screw you Corel for fucking up a great platform). So that's why I want to do something similar to Gravit Designer, universal and all, works in browsers, without the forced paywall in my face.
I always forsee premium as a way to add features to products when you need it, and as a way to let a developer know they're doing a good job. I want to make this kind of model stick, and it seems with the money-hungry fuckwits of Corel on Gravit, it ain't happening.
So that's why I want to build an designer app that does what Gravit does, except, I want premium to be more ethical, giving all the core features, add more customizabilty to the interface, and actually make the designer workbench yours. Heck, if its possible we can have Google Drive/Nextcloud integration as well for those who want cloud saves.
I badly want to do this because I believe someone out there shares my sympathy. Gravit was a nice product but was ruined by Corel's greedy paywall system. I won't be paying 99 monthly just to get offlline mode. Affinity and Figma's model is better.
Corel you fucking suck1
I'm following this fucking tutorial (https://blog.ssdnodes.com/blog/...) and everything goes well, I have docker running, docker compose installed properly, but when I start trying to create the docker-compose.yml and accessing the stupid site using the virtual host domain i set I can't it keep getting "503 Service Temporarily Unavailable" or "502 Bad gateway" what the hell am i doing wrong, I just want to get this working in my VM so i can move it to my damn server and have my own fucking cloud. This damn bullshit is exactly why i went into programming rather than dealing with configuring servers and bullshit like this i know it's outside my level of understanding but I really fucking want my own cloud system but I want it containerized for both isolation and learning purposes.
I have no idea what the hell i'm doing wrong and all the damn articles and links i'm reading aren't helping at all with my level of stupid not allowing me to understand what i'm doing wrong1
Setting up my HTPC allowed me to learn how to set up a nextcloud server. Time to start saying goodbye to Google!
Not a rant more like a question
I am currently planning to purchase a small home server + media client (with Kodi).
A small Linux Distro running the Hometheateroftware Kodi will run on the media clients (Odroid C2). The control is then over an app over the local network. The database of Kodi should be on the server in the form of a MySQL database. The movies, pictures, music are also streamed by the server (max. 2 simultaneously) via SMB (simplest variant). In addition, the server is to be accessible to the outside via a web interface to act as a cloud (maybe nextcloud). The whole should be optimized for stability and longevity. In addition, a small GitLab CE instance will probably run on the server. Do you have any comments or objections? The fact that I only take 2x ne 2 TB hard drive has the simple reason that I currently have no need for more space. Sometimes it happens to me that I forget completely obvious things :D
After 1 year decided to install Nextcloud on my Digital Ocean droplet under a subdomain. I'm happy with the results, and now I'm moving my Google Drive data (including contacts) to my Nextcloud instance.
Wish there is a Google Docs equivalent for Nextcloud I could install!.
I'm studying the chance to offer cloud space to my family for free too.3
I'm hoping to eventually set up my own server where I can store files/passwords. Does anyone have experience with Sandstorm.io, Nextcloud, or ownCloud? Are there any other alternatives you would recommend?3
Call me old fashioned, but I still use Owncloud and I intend to do. Nextcloud feels to much "hip'n'cool" and that does not always end well.
I have a small NUC-like machine in my home with an old external hdd connected to it. I use it to run my local gitlab, nextcloud and to test a few websites I build for the lolz.
If you too have a homelab, whether it's a single raspberry or an entire room full or racks, you know damn well that everything you have running locally as a web service keeps going until it doesn't, for whatever fucking reason. This time, it was the turn of my nextcloud.
The machine has arch linux running, I chose it since I already use it on my coding laptop and being a rolling release means I don't have to manually upgrade to a newer version, risking various fuck-ups and consequent screaming of profanity.
The downside is that arch is a bleeding-edge distro, so, despite being pretty good for what concerns security, as updates are pushed out some packages may still require legacy software to work as intended, since obviously not all developers for all packages can release simultaneously.
The problem was that php reached 8.2.x but nextcloud couldn't use anything beyond 8.1, so the highlighted solution was to download php-legacy, a package with a set of utilities which the cloud could use instead of mainline php.
Pretty easy, right? fuck my life, here we go.
I edited apache-httpd's configurations to link the new libraries, updated every reference in every virtual host that could possibly screw up the web server.
Then I went on and disabled the php-fpm mainline, creating a new systemd unit that would instead run the legacy executable and afterwards I edited nextcloud's additional configs so they use that instead.
Done, getting a bit dizzy, but I reboot everything and breathe.
At this point the migration should be complete, but wait, the server returns an error saying that the application is still trying to use php 8.2+...wait, what in the sysadmin Christ?
Back to nextcloud config, everything is set, everything else in every other fucking php-legacy and web server is fine, the old fpm service is disabled, I am confused, and why in the FUCKING FUCK is the new php-fpm unit failing to start at boot with "error 78/config - directory not found"? Hello? Am I being trolled by a shitty dual-core amazon fake NUC?
Maybe yes, cause it turns out that the unit was referencing a directory in the external hdd, which gets mounted at boot time after the unit itself starts, so nothing much, just a matter of tinkering with cron jobs, a reboot and at least this one is off my balls.
But why still isn't the server responding correctly? why? WHY?
After slamming my cock on the keyboard here and there scrolling back through all the config files I think to myself, hmmm, my gitlab is working flawlessly, well yeah, I didn't need to install the whole web stack, everything was nice and easy wrapped in a docker container...so why am I even here, why the fuck am I bothering with all this layered web-app bullshit, why don't I just run the up-to-date docker image that someone else has already set up for me, back up all the data and reupload them on the application?
Oh joy, you can't imagine, after 3...almost 4 hours of pure computer-touching the relief I had from seeing the blue web page with the "welcome to nextcloud" title.
Right now it's copying back all the files, and the external hdd is now linked to include the data folder.
Like really, everything was solved in two lines of bash.
I am still fuming, but at least I learned a valuable lesson, if you want a service up for yourself, implement it and deploy it as fucking easy straight-forward as you can, giving MAXIMUM priority to already fully-working options that are out there just waiting to be downloaded and used. I swing my scrotal sack on web-apps elegance as long as it's MY homelab in MY place.
Eat a fat dick php.
sudo pacman -Rns nextcloud
sudo systemctl disable --now php-fpm-legacy
sudo pacman -Rns php-legacy
sudo pacman -Rns $(sudo pacman -Qdtq)2
While planning my (personal) server I just seem to pile up more and more things to do/consider. Basically, for now I just want to have rclone, nextcloud and jellyfin, plus some usenet stuff later on. But I want to have the whole installation and configuration automated as far as possible, since I'll at first it will run in a test environment and needs to be migrated to another server at a point, possibly even another OS. So I suppose that means docker, docker-compose and Chef (any better options?). I want SSL: Traefik. User management / auth? RADIUS, LDAP. SSO? keycloak. I also need to deal with virtual hosts. And probably much more..
Since I just have basic Linux knowledge and have no real experience with any of the other technologies, I feel a bit lost. I just got to the abovementioned software due to some ddg research. I don't mind digging deep, I want to learn (which is half the reason for this project), but it's not easy to the the best way to set this up.11
i need an adult. I know noone who would understand my worries, so you guys need to be it.
i have a nextcloud running on my raspberry pi. performance is horrible, dont ask, but it works.
i mostly use it to backup the photos of my phone sd card every night when my phone charges. Internally this works good. If i am elseplace it wont for obvious reasons.
In my youthful joy of doom i opened port 443 and forward it to my raspi. I get internet via cable and my ip is pretty much static (it was the same for 10 months). So external access is provided.
Now i thought, its stupid that i cannot sign an ssl certificate cause i dont have a domain. Lets buy domain. But before i do that i did some try runs with duckdns to test the principle.
Some back and forth, it works now. Pretty god, i could even make a cron job on the raspbi to renew (that should work right?). Only problem. randoname.duckdns.org doesnt work internally. Or should not at least.
So i googled a bit and it turns out that my router (a cable fritz!box i bought myself) can be a local network dns. Or cannot. Regardless what i try, it doesnt accept the changed config file.
Now the problem.
It works anyway. randoname.duckdns.org points to my external "static" ip and resolves to that from my internal network..so it works on my phone or laptop. if i traceroute the thing it goes via two hops out and finishes in less than 1ms.
Now to the problem:
I have no fokkin clue why. The expected behaviour would be that it shouldnt work. If i do what i intended todo on pc in the hosts file tracert works correctly, directly pointing to the internal ip.
What i cannot figure out, is it the fritz!box being smart? Is it my ISP being smart?
Reason to rant: i have absolutly NOONE to ask, i know not a single person who would even understand what troubles me. I want to learn, i want to know WHY not just some mindless russian patchwork of "if it works its good enough".
Just went to update my nextcloud instance, is there an archive of packages for archlinux ARM, nextcloud stable isnt compatible with php 7.2.
I regularly clear /var/cache/pacman/pkg yes i already checked...
I'm planning to switch from ownCloud to Nextcloud in the next weeks. Does anyone have experience with that or maybe some advice? Aside from creating Backups.7
So I have a raspberry pi running nextcloud and I want to add a second one to improve the availability.
I'm new to HA clusters. What stack would you recommend?17
Just wanted to do some scripted image resizing for school in school because the teacher asked me to help her with that.
So I thought: Let's just write a tiny script. Written the script in almost no time (just iterates over all jpg's and resizes them)
Now I tried to run it. Didn't have my laptop so I had to somehow run it on their windows PCs. At least it's windows 10, unlike other schools that still run XP and stuff so I thought it might be doable. Well guess what, nope it wasn't.
First tried to install imagemagick, that didn't work as only teacher accounts have admin and the teacher was already pretty scarred once he saw me doing stuff in powershell so I thought I'd better not ask to do this via a teacher account and mess with stuff as admin.
Next method: Installing msys2. That worked at least (after taking forever to install and having to mess with the av software to get it to run).
And there comes the next problem: pacman doesn't connect via the proxy so I can't download any packages. There is free wifi but only for teachers, and students aren't going to get access until the school finally has a faster connection because they'd (understandably) cause this connection to be constantly overloaded. I just happen to have access to this wifi network, too, because at least the guys from the IT dept know how bad using proxies under linux is. So I connect via wifi and it works. At least I thought: After running the script it yields weird errors about unsupported arguments even though the command is exactly the same I have been using for years (already checked typos twice)
Then got the idea of simply installing imagemagick on termux on android and transferring the files onto my phone.
Too bad we aren't allowed to attach our own USBs to the pcs. Luckily I got a rooted phone so I simply activate adb over network and connect to it.
After downloading the platform-tools I can't run them because of AV software. Luckily there is an option to add an exception per executable so I do that. After doing that it works.... nope it doesn't. The wifi only allows 443/tcp and 80/tcp, even for internal network devices.
So that's it. I'm simply going to upload that stuff to my nextcloud and convert it at home.
Windows, I hate you!!!2
Hi everyone, I have a question about VPN and hosting.
I have rpi which runs ubuntu where I have several things running like nextcloud, transmission, minidlna, samba etc.
I want to use a VPN due to torrenting via transmission on the pi. I had used private internet access(PIA) before and I'm thinking to go back to them as I had issue only once with them.
Question is if I had installed their client and connected to VPN, would I still be able to access to my services over the internet? As per my understanding only the outgoing and incoming generated from outgoing should follow the VPN tunnel, therefore interacting with my pi with it's public IP should still be possible, am I right?
I'm a newb when it comes to web stuff so any help is appreciated, also you can recommend other VPN providers if you think PIA sucks for any reason.4
Oh, there are hundreds that I've started categorizing them. They outgrew the storage capacity of my head / brain. I've tried a lot of productivity tools to organize them, but in the end, all my project ideas just remain ideas scattered somewhere unless I see it action and go like, "Hey, I had that same idea. I wonder when they got the idea. Was it before or after I had it?" In the end, I just console myself saying that for me it was only an idea in my head when those people saw it through execution and has a working product. The next step for me is to get along with them and collaborate and make that idea better rather than re-invent the same wheel again according to my idea.
Nextcloud is the biggest example of an idea that came to me and remained in my head and is still on a todo list somewhere.
Note: Do not work on projects off of cloud on multiple computers. You will most likely start losing code. As I did.
Was confident in my cloud setup until this happened, so began using git for projects that I actually care about.
Does anyone have experience with Google Drive (GSuite) and rclone? I want to use it as a storage for jellyfin (emby fork) and Nextcloud, with the first being only saved there and the latter either as or with a backup.
Need an advice on syncing large amounts of small files. For starters I have 2 million of ~20-25kb sized files in dropbox. Syncing it with my local machine is a pain in the ass (API limits syncing to 4-5 files per second, so I will need at least 96 hours just to take out everything from dropbox).
Recently I started looking into self hosted cloud solutions. I fiddled around with nextcloud but it's client limits syncing to 5-6 files per second as well. Accessing nextcloud via webdav from a rclone client I was able to up the syncing speed to 15-25 files per second but that's all.
Any solutions where I should host 2mil of files and would be able to sync them fast to my own machine with some cloud client when I need it to?5
I finally got around to setting up my own cloud with nextcloud on my own dedicated server.
Just setting up Nextcloud alone was not really the challenge ( I've set up at least 2 Nextcloud instances in the past ).
The actual challenge was to install /e/ OS on my mobile phone and get it to work with my Nextcloud instance.
It's not all performant, buttery-smooth or super-fast yet, but for a one-person / user-cloud, I think it should be just fine.
There's still room for improvement in terms of server-side performance, but it's working fine with the basics at least.
I need to figure / iron out some issues like social federation via ActivityPub not working, Nextcloud SMS not syncing up my SMS, Mail app crashing because I used a self-hosted Nextcloud instance, etc; but those are things I could work on slowly, in the course of time.
No, the server is not physically controlled by me, yet ( it's a dedicated box server though. Still, hosted and physically controlled by a provider ).
I intend on setting up another 'replica' on a RaspberryPi which I will then make primary, connecting to the internet via DynamicDNS.
I'll probably keep the server as a fallback / backup server just in case my home server loses connectivity.
Taking back control from Big Tech is something I intend on pursuing actively this year. I've had the idea in my head for too long that it has started to fester.
This is only a first step, of many, that needs to follow, in order for me to take control back from Big Tech.
Yes, there still is some room for improvement, but I think for now ‒
Someone posted a link to a 30-day-security-challenge here on devRant some time ago and I just thought well, why not try to migrate away from the big companies - I've been using OneDrive as my only cloudstorage since the time when it was called SkyDrive and I've been hosting my Emails at outlook (via Live Custom Domains, a service that does not even exist anymore) for about 8 years now. Since I've always been lazy and since exchange activesync is a great feature if you have multiple calendars and want to sync them and your contacts to several devices I never tried to switch but now I am half done with migrating my data to my own nextcloud installation and my emails to my own mail server - since I don't want to loose the exchange functionality I am also setting up Z-Push and oh boy, this thing is bitching around but my webmail is already nicely integrated into nextcloud, IMAP / SMTP is up, configured and secured (still have to mess around with spamassassin as this email adress is floating around the web for about 10 years now). The only things to do is to get Z-Push work with STARTTLS and the card/caldav backend running and then the basic setup should be done.
I am just wondering if someone could hand me over a guide on how to sign / encrypt emails (GPG?)