Technically I don’t think any Greek layout uses a different Unicode codepoint for the question mark. In fact, the ordinary semicolon symbol is used, so what the meme describes would probably not happen IRL.
Does all this make it any less funnier? No. It’s still brilliant.
In Unicode, it is separately encoded as U+037E ; GREEK QUESTION MARK, but the similarity is so great that the code point is normalised to U+003B ; SEMICOLON, making the marks identical in practice.
I’m still curious whether it would be accepted by the code interpreters / compilers of various languages. I’m not bold enough to assume they all normalise properly.
Unicode should have enforced the principle of using the same encoding for similar looking characters like they did with CJK instead of allowing bullshit like the Cyrillic “o” or the Greek question mark.
ForgeFed is an ActivityPub Extension to allow cross-forge (git server) issues, pull requests etc. without having to create an account on each server. Forgejo (a Gitea soft-fork) is actively working on this integration.
Having this feature would be very useful. Many big open source projects run their own gitlab instance, which add extra frictions for contributing because you need yet another gitlab account just for those projects.
~95% of the JS code you see on the Web has semicolons. Apparently, a lot of programmers think it’s worth that extra keystroke to avoid these types of bugs. I agree with them. The difficulty with programming isn’t “Arrgh, there are too many keystrokes, my hands are tired!” It’s “Arrgh, HTF did this bug get in here?!?”
They made it a named function, but this is literally how you format anonymous functions in js, a key feature that usually gets called with stuff like an onclick or onload call in the html domain.
I use prettier which by default adds semicolons. Coming from predominately doing backend stuff (mostly in Java) I don’t really mind, especially when the formatter adds them for me
Something similar happened to me a while back. I was copying some code from a Mac to a remote Linux host. For some reason the Mac was using a thing called an “en dash” – which is slightly longer than a regular hyphen - and was really fucking frustrating to figure out.
I don’t know why I’m here commenting about this, but I love type, so:
Hyphen (-): the short one, used for hyphenated words. fire-eaters. Close-up.
en-dash (–): slightly longer, traditionally the length of a lowercase"n" in the typeface. Used between for things like a timeframe. 10–11:30, August–October
em-dash (—): the longest of the three, and the length of a lowercase “m”. Used as a punctuation mark to denote a side comment or to abruptly cut off a sentence. “It’s a great punctuation mark—in fact I overuse it—but it’s still useful.” “Hey where are you going with that giant—”
I didn’t bother to double check the definitions, so there might be more specific rules, but these are my rules of thumb.
Some mac apps have some quirks, the default note app was probably not meant for pasting code in, but when you do it changes the quotes and makes them all fancy. Drives me up the wall and there’s nobody to blame but me.
I was looking for this. Some text from webpages end up pasting that way too, even on non-mac systems, and it is utterly infuriating. Nothing I hate more than having to paste something into notepad++ so I can fix all the stupid quotes from some online tutorial that is giving you things to paste into a command prompt.
Ruby’s popularity in the early 10s thanks to Ruby on Rails feels like it happened by accident. The language is hard to read and low performance, but Rails is completely automagic. But this is also the worst thing about rails. You create your app fast, but then maintaining it is expensive because you can’t onboard new developers easily. Even if they’re familiar with rails’ automagicisms, it will take them quite some time to parse what the hell the code is doing.
Meanwhile I seem to recall Ruby’s creator finding the situation of his language being popular because he’d created it as an experiment and never thought it would be used in production grade environments
<span style="color:#323232;">for item in array do
</span><span style="color:#323232;"> puts item[:name]
</span><span style="color:#323232;">end
</span>
Whew, iterating and working with data in Ruby is so hard. How does anyone read this stuff.
low performance
Ruby is a syntax-sugar-loaded C-wrapper, just like Python and countless other languages that don’t compile straight to machine code. If anything other than C and Rust are slow to you, then sure, maybe Ruby isn’t a good fit for your project (but Crystal might be).
create your app fast
Damn right, I’m two or three times as productive as I ever was in C#/Razor, Java/Spring or kludging through the countless JS boilerplate-heavy web frameworks.
but then maintaining it is expensive
As with any app that grows into something successful and widely used, technical complexity becomes exponential. I’ve found once web applications grow to a certain number of models and controllers, the relationships between them start to grow exponentially as well. This means one small change can ripple throughout your application and have unintended consequences where you least expect.
This is not even remotely a unique problem to Ruby. It’s happened across every project I’ve seen that grows beyond 30 models and a couple of dozen controllers, regardless of language. This is why unit testing is so important.
But, specifically you mentioned you can’t “onboard new developers easily”. I don’t see how. I’ve taken two CS grads straight out of college and had them adding features with tests within a couple of days on Ruby projects. Ruby was designed to be most friendly to humans, not the compiler. If Rails is what is tripping you up, imagine trying to learn a new web framework on top of an even more complicated language than Ruby. I just don’t see this argument at all, from my experiences.
Ruby’s creator finding the situation of his language being popular because he’d created it as an experiment
Pretty sure most any language that was created by an individual and not by BigCorp™ is a feat in and of itself. This speaks more widely to a language’s capabilities and value if it can reach popularity without corporate backing. This argument seems to imply that because of it’s origin, it will always be some kind of experimental toy that was never intended for wide-use.
Meanwhile, Linus Torvalds:
I’m doing a (free) operating system (just a hobby, won’t be big and professional like gnu) for 386(486) AT clones.
Things have to start somewhere, I guess?
I kindly ask you to be more constructive in your criticism of Ruby. It’s a great, powerful language with a low barrier to entry. There’s no reason to spread FUD about it.
First time I’ve ever heard someone call a for loop “weird“. They’ve been around for 50 years 😂
The whole point was on readability, not trying to make rubocop be quiet. Sure, .each is great, but I’m not sure about it being shorthand. What did you save? Like 3 characters? I find the for loop more readable unless I’m method chaining.
Not in ruby, the for loop was initially put there to make it friendly for people from other languages and is discouraged. It’s just syntax sugar on top of each.
It’s valid syntax, it’s part of Ruby. It’s easy to read and familiar across many languages. Write what you want to write, I’m not sure why you feel the need to finger wag.
Was about to say as someone who’s been using Ruby for over a decade, 8 of which professionally, I’ve never once come across a for loop. each on the other hand, all day every day.
Yeah there’s multiple people in here saying that and it seems like maybe they’ve never actually written Ruby. I don’t think Ruby is a good language for writing business apps in, but it’s incredibly easy to read. Way easier than pretty much every single language out there.
Now if you start including shit like rails, sure. But that’s not Ruby. That a framework and just like Spring or Django or Boost or whatever, it has its own semantics and can be incredibly difficult to read. That has nothing to do with the language though.
Which benchmark are you talking about? The most common I know of is the computer language benchmarks game, here’s a nice implementation of it.. You can see ruby is actually a decent bit faster.
It’s an annoyingly persistent myth that ruby is significantly slower than python.
He presented the issue with gitlab very well. Setting up an entire new account is the major reason (besides time) I don’t contribute to projects on other gitlab instances. For some reason Gitlab management didn’t think it important at all (maybe even considered it a feature).
Their documentation has been frustratingly outdated at times too. But since GitHub is MS owned there are better options. I prefer codeberg for having an actual account on.
Yeah, been using GitLab semi-professionally for a while and have accounts on multiple instances, it never asked me to have a credit card on file, and I just don’t put in a phone number. Saying that it requires it is sensationalism.
I guess Canada also doesn’t need credit cards. But as you can see from this comment, there are countries in which credit cards are required: lemmy.one/comment/3041845
Definitely not the case. I’ve been doing C# for over 20 years. For the last 6 years I’ve used it to write Linux services exclusively.
The “Java” relation was true 20 years ago, and the “Linux” argument was true 7 years ago. But neither hold any water anymore. It’s a great language and framework to write a wide variety of software with.
programmer_humor
Active
This magazine is from a federated server and may be incomplete. Browse more on the original instance.