It looks like AssDB uses a weird SQL syntax? Is it worth upgrading to, I hear it’s great at pulling information out of unstructured and even imaginary data sources?
FTP isn’t really used much any more. SFTP (file transfers over SSH) mostly took over, and people that want to sync a whole directory to the server usually use rsync these days.
Isn’t Cyberduck a paid program though? I remember trying it, but I can’t remember why I went back to filezilla. I thought it was because my trial for Cyberduck expired.
I fell like the reason nobody uses FileZila and etc anymore is because everybody that wanted it migrated to Linux already. So seriously, it already happened.
Somehow I miss those days. Now you need weeks of training to understand the black magic behind all the build/deployment stuff in whatever cloud provider your company decided to use…
We got our own platform based on kubernetes and cncf stuff and we don’t have to care anymore about the metal underneath. AWS? OTC? Azure? Thats just a target parameter, platform does the rest. It’s great.
How often do you switch cloud providers that this is even a real rather than a hypothetical benefit? (Compared to the cost of dealing with a much more complicated stack.)
I manage a stack like this, we have dedicated hardware running a steady state of backend processing, but scale into AWS if there’s a surge in realtime processing needed and we don’t have the hardware. We also had an outage in our on prem datacenter once which was expensive for us (I assume an insurance claim was made), but scaling to AWS was almost automatic, and the impact was minimal for a full datacenter outage.
If we wanted to optimize even more, I’m sure we could scale into Azure depending on server costs when spot pricing is higher in AWS. The moral of the story is to not get too locked into any one provider and utilize some of the abstraction layers so that AWS, Azure, etc are just targets that you can shop around for by default, without having to scramble.
The best Hello World I saw used a random library. Because there’s no true random without hardware, the author figured out the correct seed to write Hello World with “random” characters. I’ve used that to show junior devs that random in programming doesn’t mean truly random.
One could argue the requirements have changed because the security and compliance part of the world finally caught up to modern software delivery concepts. Even the most dinosaur apps at compliant orgs are being dragged kicking and screaming into new CI/CD tools where applying governance and custody chains and permissions and approvals are all self documented automated hooks.
Here’s the core issue. The developer didn’t know his rights, and made a mistake. I’m not criticizing, people make a career dealing with crap like this. But if you want to make a business out of something, it’s worth it to do some research or talk to a lawyer. I believe the MIT license has its place but, from what the OP said, this isn’t it.
I did not want to make a business out of this library. I don’t want money for it.
All I would’ve wanted is that the people at Apple would’ve given me a heads up beforehand, so I would’ve been prepared for it and not caught on surprise. And a that they do a version upgrade when I release a new bugfix release.
This is not a license issue. I was well aware of the consequences when I chose the MIT license. This is not about money.
You specifically said you chose the MIT license because you wanted to use it in commercial projects. That’s business, no matter how small. As the owner of the property, you could have used any and all licenses available to you. Also, if you wanted to require users of your code to attribute or notify you, you could have. If you want to be disappointed in their behavior that’s perfectly fine, too. Corporations usually disappoint if you have any altruistic expectations of them.
In this regard you are right. I could’ve chosen AGPL and use it in my commercial project nonetheless. I wasn’t aware of that at the time, and that was a mistake.
That said, I don’t expect all users to notify me. But if a company like Apple, with millions of users, exposes me to even a fraction of its users - then yes. I expect a mail beforehand. I did not sign up for this.
Agreed. Free licenses should NEVER be applied to Apple-specific tools. They don’t want to help the FOSS community, so we shouldn’t help them back. Make them pay for it, or make them make their own version.
You can use your library for commercial projects that you have. Just have dual license that requires payment for commercial use or something similar. You don’t have to pay yourself
I think that’s why Github suggests MIT as default. Unaware people will just put that. Most open source people just code things they want without thinking much on other aspects. We really need some sort of enforcement to stop companies banking on voluntary work done for the community.
It’s probably a single dev that made the decision, then moves onto something else. They (probably?) don’t have the ability to just raise a recurring PO etc to easily pay you and don’t care enough to worth through the paperwork.
If you had a paid licencing model they may have done it, or just found another lib/ wrote their own.
I suppose in the days of ‘Cloud Hosting’ a lot of people (hopefully) don’t just randomly upload new files (manually) on a server anymore.
Even if you still just use normal servers that behave like this, a better practice would be to have a build server that creates builds, like whenever you check code into the Main branch, it’ll create a deploy for the server, and you deploy it from there - instead of compiling locally, opening filezilla and doing an upload.
If you’re using ‘Cloud Hosting’ - for example AWS - If you use VMs or bare metal - you’d maybe create Elastic Beanstalk images and upload a new Application or Machine Image as a new version, and deploy that in a more managed way. Or if you’re using Docker, you just upload a new Docker image into a Docker registry and deploy those.
For some of my sites, I still build on my PC and rsync the build directory across. I’ve been meaning to set up Gitlab or something similar and configure automated deployments.
This is what I do because my sites aren’t complicated enough to warrant a build system. Personally I think most websites out there are over-engineered. Example: a Discord friend made a React site that displays stats from a gaming server. It looks nice, but you literally can’t hyperlink to any of the data, it can only be loaded dynamically and only looks coherent on a phone in portrait mode. There are a lot of people following trends (some good trends) but without really thinking about why.
I’m starting to like the htmx model a lot. Server-rendered app that uses HTML attributes to configure the dynamic bits (e.g. which URL to hit and which DOM element to insert the response into). Don’t have to write much JS (or any in some cases).
you literally can’t hyperlink to any of the data
I thought most React-powered frameworks use a URL router out-of-the-box these days? The developer does need to have a rough idea what they’re doing, though.
Yea, I wasn’t saying it’s always bad in every scenario - but we used to have this kinda deployment in a professional company. It’s pretty bad if this is still how you’re doing it like this in an enterprise scenarios.
But for a personal project, it’s alrightish. But yea, there are easier setups. For example configuring an automated deployed from Github/Gitlab. You can check out other peoples’ deployment config, since all that stuff is part of the repos, in the .github folder. So probably all you have to do is find a project that’s similar to yours, like “static file upload for an sftp” - and copypaste the script to your own repo.
They have bundled malware from the main downloads on their own site multiple times over the years, and even denied it and tried gaslighting people that AVs were giving false positives because AV companies are paid off by other corporations. And the admin will even try to delete the threads about this stuff but web archive to the rescue…
You know what? I didn’t believe you, since I’m using it for a long time on Linux and never had any issues with it. Today, when I helped a friend (on Windows) with some SFTP transfer and recommended FileZilla was the first time I realised the official Downloads page provides Adware. The executable even gets flagged by Microsoft Defender and VirusTotal. That’s actually REALLY bad. Isn’t FileZilla operated by Mozilla? Should I stop using it, even though the Linux versions don’t have sketchy stuff? It definitely leaves a really bad taste.
Yeah, it’s bad. Surprised they’re still serving that crap in their own bundle but i guess some things don’t change.
Filezilla is no relation to mozilla. But yeah i moved away from it years ago. The general recommendation I’ve seen is “anything but filezilla”. Personally i use winscp for windows, and will have to figure out what to use when i switch my daily driver to Linux.
programmer_humor
Newest
This magazine is from a federated server and may be incomplete. Browse more on the original instance.