There have been multiple accounts created with the sole purpose of posting advertisement posts or replies containing unsolicited advertising.

Accounts which solely post advertisements, or persistently post them may be terminated.

SirGolan

@[email protected]

This profile is from a federated server and may be incomplete. Browse more on the original instance.

SirGolan ,

Man that video irks me. She is conflating AI with AGI. I think a lot of people are watching that video and spouting out what she says as fact. Yet her basic assertion is incorrect because she isn’t using the right terminology. If she explained that up front, the video would be way more accurate. She almost goes there but stops short. I would also accept her saying that her definition of AI is anything a human can do that a computer currently can’t. I’m not a fan of that definition but it has been widely used for decades. I much prefer delineating AI vs AGI. Anyway this is the first time I watched the video and it explains a lot of the confidently wrong comments on AI I’ve seen lately. Also please don’t take your AI information from an astrophysicist, even if they use AI at work. Get it from an expert in the field.

Anyway, ChatGPT is AI. It is not AGI though per recent papers, it is getting closer.

For anyone who doesn’t know the abbreviation, AGI is Artificial General Intelligence or human level intelligence in a machine. ASI is Artificial Super Intelligence which is beyond human level and the really scary stuff in movies.

SirGolan ,

It’s interesting. I’ve been seeing a lot of the incorrect ideas from this video being spread around lately, and I think this is the source. I’m surprised there aren’t more people correcting the errors, but here’s one from someone in the banking industry who completely refutes her claims of not being able to use AI to approve mortgages. If I had more time, I’d write up something going over all the issues in that video. Like she even misunderstands how art works unrelated to AI. She is basically saying that anything she doesn’t like isn’t art. That’s not how that works at all. Anyway, it’s really hard to watch that video as someone who works in the field and has a much better understanding of what she’s talking about than she does. I’m sure she knows a lot more about astrophysics than I do. She also made a video saying all humanoid robots are junk. She’s very opinionated about things she doesn’t have experience with, which again, is her right. Just I think a lot of people put weight into what she says and her opinions because she’s got a PhD after her name. Doesn’t matter that it’s not in AI or robotics.

SirGolan ,

I think it might require plus but the iOS And Android apps do support voice only conversation. You have to go into beta features and enable it.

SirGolan ,

Check out this recent paper that finds some evidence that LLMs aren’t just stochastic parrots. They actually develop internal models of things.

SirGolan ,

Wikipedia: In copyright law, a derivative work is an expressive creation that includes major copyrightable elements of a first, previously created original work.

I think you may be off a bit on what a derivative work is. I don’t see LLMs spouting out major copyrightable elements of books. They can give a summary sure, but Cliff Notes would like to have a word if you think that’s copyright infringement.

SirGolan ,

From what I’ve seen, here’s what happened. GPT 4 came out, and it can pass the bar exam and medical boards. Then more recently some studies came out. Some of them from before GPT 4 was released that just finally got out or picked up by the press, others that were poorly done or used GPT 3 (probably because of gpt 4 being expensive) and the press doesn’t pick up on the difference. Gpt 4 is really good and has lots of uses. Gpt 3 has many uses as well but is definitely way more prone to hallucinating.

ChatGPT generates cancer treatment plans that are full of errors — Study finds that ChatGPT provided false information when asked to design cancer treatment plans (www.businessinsider.com)

ChatGPT generates cancer treatment plans that are full of errors — Study finds that ChatGPT provided false information when asked to design cancer treatment plans::Researchers at Brigham and Women’s Hospital found that cancer treatment plans generated by OpenAI’s revolutionary chatbot were full of errors.

SirGolan ,

What’s with all the hit jobs on ChatGPT?

Prompts were input to the GPT-3.5-turbo-0301 model via the ChatGPT (OpenAI) interface.

This is the second paper I’ve seen recently to complain ChatGPT is crap and be using GPT3.5. There is a world of difference between 3.5 and 4. Unfortunately news sites aren’t savvy enough to pick up on that and just run with “ChatGPT sucks!” Also it’s not even ChatGPT if they’re using that model. The paper is wrong (or it’s old) because there’s no way to use that model in the ChatGPT interface. I don’t think there ever was either. It was probably ChatGPT 0301 or something which is (afaik) slightly different.

Anyway, tldr, paper is similar to “I tried running Diablo 4 on my Windows 95 computer and it didn’t work. Surprised Pikachu!”

SirGolan ,

I agree, but only because they used GPT 3.5 and not 4. Not that I think 4 would have been perfect or that you should follow medical advice from LLMs right now, but it would have been much more accurate.

SirGolan ,

Hah! That’s the response I always give! I’m not saying our brains work the exact same way because they don’t and there’s still a lot missing from current AI but I’ve definitely noticed that at least for myself, I do just predict the next word when I’m talking or writing (with some extra constraints). But even with LLMs there’s more going on then that since the attention mechanism allows it to consider parts of the prompt and what it’s already written as it’s trying to come up with the next word. On the other hand, I can go back and correct mistakes I make while writing and LLMs can’t do that…it’s just a linear stream.

SirGolan ,

Yeah what’s interesting is it was just published this week even though they did the tests in April. April is like 100 AI years ago.

SirGolan ,

extraordinary claims without extraordinary proof

What are you looking for here? Do you want it to be self aware and anything less than that is hot garbage? That latest advances in AI have many uses. Sure Bitcoin was over hyped and so is AI, but Bitcoin was always a solution with no problem. AI (as in AGI) offers literally a solution to all problems (or maybe the end of humans but hopefully not hah). The current tech though is widely useful. With GPT4 and GitHub Copilot, I can write good working code at multiple times my normal speed. It’s not going to replace me as an engineer yet, but it can enhance my productivity by a huge amount. I’ve heard similar from many others in different jobs.

SirGolan ,

As I see it, anybody who is not skeptical towards “yet another ‘world changing’ claim from the usual types” is either dumb as a doorknob, young and naive or a greedy fucker invested in it trying to make money out of any “suckers” that jump into that hype train.

I’ve been working on AI projects on and off for about 30 years now. Honestly, for most of that time I didn’t think neural nets were the way to go, so when LLMs and transformers got popular, I was super skeptical. After learning the architecture and using them myself, I’m convinced they’re part of but not the whole solution to AGI. As they are now, yes, they are world changing. They’re capable of improving productivity in a wide range of industries. That seems pretty world changing to me. There are already products out there proving this (GitHub Copilot, jasper, even ChatGPT). You’re welcome to downplay it and be skeptical, but I’d highly recommend giving it an honest try. If you’re right then you’ll have more to back up your opinion, and if you’re wrong, you’ll have learned to use the tech and won’t be left behind.

SirGolan ,

Yeah, I generally agree there. And you’re right. Nobody knows if they’ll really be the starting point for AGI because nobody knows how to make AGI.

In terms of usefulness, I do use it for knowledge retrieval and have a very good success rate with that. Yes, I have to double check certain things to make sure it didn’t make them up, but on the whole, GPT4 is right a large percentage of the times. Just yesterday I’d been Googling to find a specific law or regulation on whether airlines were required to refund passengers. I spent half an hour with no luck. ChatGPT with GPT4 pointed me to the exact document down to the right subsection on the first try. If you try that with GPT3.5 or really anything else out there, there’s a much higher rate of failure, and I suspect a lot of people who use the “it gets stuff wrong” argument probably haven’t spent much time with GPT4. Not saying it’s perfect-- it still confidently says incorrect things and will even double down if you press it, but 4 is really impressive.

Edit: Also agree, anyone saying LLMs are AGI or sentient or whatever doesn’t understand how they work.

SirGolan ,

Wait a second here… I skimmed the paper and GitHub and didn’t find an answer to a very important question: is this GPT3.5 or 4? There’s a huge difference in code quality between the two and either they made a giant accidental omission or they are being intentionally misleading. Please correct me if I missed where they specified that. I’m assuming they were using GPT3.5, so yeah those results would be as expected. On the HumanEval benchmark, GPT4 gets 67% and that goes up to 90% with reflexion prompting. GPT3.5 gets 48.1%, which is exactly what this paper is saying. (source).

SirGolan ,

Hmm that’s incorrect. ChatGPT (if you pay for it) does both.

SirGolan ,

If we are talking Copilot then that’s not ChatGPT. But I agree it’s ok. Like it can do simple things well but I go to GPT 4 for the hard stuff. (Or my own brain haha)

SirGolan ,

Yes available to anyone in the API or anyone who pays for ChatGPT subscription.

SirGolan ,

Oh ok! Got it. I read it as you saying ChatGPT doesn’t use GPT 4. It’s still unclear what they used for part of it because of the bit before the part you quoted:

For each of the 517 SO questions, the first two authors manually used the SO question’s title, body, and tags to form one question prompt3 and fed that to the Chat Interface [45] of ChatGPT.

It doesn’t say if it’s 4 or 3.5, but I’m going to assume 3.5. Anyway, in the end they got the same result for GPT 3.5 that it gets on HumanEval, which isn’t anything interesting. Also, GPT 4 is much better, so I’m not really sure what the point is. Their stuff on the analysis of the language used in the questions was pretty interesting though.

Also, thanks for finding their mention of 3.5. I missed that in my skim through obviously.

SirGolan ,

Yeah I think you’re right on about the students not being able to afford GPT4 (I don’t blame them. The API version gets expensive quick). I agree though that it doesn’t seem super well put together.

SirGolan ,

I’ve been making the same or similar arguments you are here in a lot of places. I use LLMs every day for my job, and it’s quite clear that beyond a certain scale, there’s definitely more going on than “fancy autocomplete.”

I’m not sure what’s up with people hating on AI all of a sudden, but there seems quite a few who are confidently giving out incorrect information. I find it most amusing when they’re doing that at the same time as bashing LLMs for also confidently giving out wrong information.

SirGolan ,

Yeah, I think that’s a big part of it. I also wonder if people are getting tired of the hype and seeing every company advertise AI enabled products (which I can sort of get because a lot of them are just dumb and obvious cash grabs).

At this point, it’s pretty clear to me that there’s going to be a shift in how the world works over the next 2 to 5 years, and people will have a choice of whether to embrace it or get left behind. I’ve estimated that for some programming tasks, I’m about 7 to 10x faster when using Copilot and ChatGPT4. I don’t see how someone who isn’t using AI could compete with that. And before anyone asks, I don’t think the error rate in the code is any higher.

SirGolan ,

The one I like to give is tool use. I can present the LLM with a problem and give it a number of tools it can use to solve the problem and it is pretty good at that. Here’s an older writeup that mentions a lot of others: www.jasonwei.net/blog/emergence

Tech experts are starting to doubt that ChatGPT and A.I. ‘hallucinations’ will ever go away: ‘This isn’t fixable’ (fortune.com)

Tech experts are starting to doubt that ChatGPT and A.I. ‘hallucinations’ will ever go away: ‘This isn’t fixable’::Experts are starting to doubt it, and even OpenAI CEO Sam Altman is a bit stumped.

SirGolan ,

If that were true, it shouldn’t hallucinate about anything that was in its training data. LLMs don’t work that way. There was a recent post with a nice simple description of how they work, but I’m not finding it. If you’re interested, there’s plenty of videos and articles describing how they work.

SirGolan ,

You guys should all check out Andrej Karpathy’s neural networks zero to hero videos. He has one on LLMs that explains all this.

Netflix lists $900,000 AI job as actors and writers continue to strike (www.engadget.com)

Netflix lists $900,000 AI job as actors and writers continue to strike::Will this pair of Hollywood strikes ever end? It looks like the big corporations are digging in for a long battle, illustrated by Netflix’s recent job posting for a machine learning platform product manager.

SirGolan ,

All the articles about this I’ve seen are missing something. Netflix has been using machine learning in a bunch of ways for quite a few years. I bet this position they’re hiring for has been around for most of that time and isn’t some new “replace all actors and writers with AI” thing. Here’s an article from 2019 talking about how they use AI. That was the oldest I could find but someone I know was working on ML at Netflix over a decade ago.

SirGolan ,

You should check out the short story Manna. It’s maybe a bit dated now but explores what could go wrong with that sort of thing.

SirGolan ,

What I wonder is why more cars don’t have HUDs that are projected onto the windshield. That tech has been around and in cars for over 25 years. You don’t have to take your eyes off the road at all.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • random
  • lifeLocal
  • goranko
  • All magazines