I've trained as a neuroscientist and written a book about consciousness. I've worked in machine learning and built products for over 20 years and now use AI a fair bit in the ed-tech work we do.
So I've seen how the field has progressed and also have been able to look at it from a perspective most AI/engineering people don't -- what does this artificial intelligence look like when compared to biological intelligence. And I must say I am absolutely astonished people don't see this as opening the flood-gates to staggeringly powerful artificial intelligence. We've run the 4-minute mile. There are hundreds of billions of dollars figuring out how to get to the next level, and it's clear we are close. Forget what the current models are doing, it is what the next big leap (most likely with some new architecture change) will bring.
In focusing on intelligence we forget that it's most likely a much easier challenge than decentralized cheap autonomy, which is what took the planet 4 billion years to figure out. Once that was done, intelligence as we recognize it took an eye-blink. Just like with powered-flight we don't need bioliogical intelligence to transform the world. Artificial intelligence that guzzles electricity, is brittle, has blind spots, but still capable of 1000 times more than the best among us is going to be here within the next decade. It's not here yet, no doubt, but I am yet to see any reasoned argument for why it is far more difficult and will take far longer. We are in for radical non-linear change.
I've worked in AI for the past 30 years and have seen enthusiasm as robust as yours go bust before. Just because some kinds of narrow AI have done extraordinarily well -- namely those tasks that recognize patterns using connections between FSMs -- does not mean that same mechanisms will continue to scale up to human-level cognition, much less exceed it any time soon.
The breakthroughs where deep AI have excelled -- object recognition in images, voice recognition and generation, and text-based info embedding and retrieval -- these require none of the multilevel abstraction that characterizes higher cognition (Kahneman's system 2 thinking). When we see steady progress on such frontiers, only then a plausible case can and should be made that the essentials of AGI are indeed within our grasp. Until then, plateauing at a higher level of pattern matching than we had expected -- which is what we have seen many times before from narrow AI -- this is not sufficient evidence that the other requisite skills needed for AGI are surely just around the corner.
So I am a neophite in this area, but my thesis for why "this time is different" compared to previous AI bubbles is that this time there exist a bunch of clear products (or paths to products) that work and only require what is currently available in terms of technology.
Coding assistants today are useful, image generation is useful, speach recognition/generation is useful.
All of these can support businesses, even in their current (early) state. Those businesses have value in funding even 1% improvements in engineering/science.
I think that this is different than before, where even in the 80s there were less defined products, amd most everything was a prototype that needed just a bit more research to be commercially viable.
Where as in the past, hopes for the technology waned and funding for research dropped off a cliff, today's stuff is useful now, and so companies will continue to spend some amount on the research side.
Really? I work in AI and my biggest concern is that I don't see any real products coming out of this space. I work closer to the models, and people in this specific area are making progress, but when I look at what's being done down stream I see nothing, save demos that don't scale beyond a few examples.
This is literally all I see right now. There's some really fun hobbyist stuff happening in the image gen area that I think is here to stay, but LLMs haven't broken out of the "autocomplete on steroids" use cases.
Can you give me examples of 5, non-coding assistant, profitable use cases for LLMs that aren't still in the "needed just a bit more research to be commercially viable" stage?
I love working in AI, think the technology is amazing, and do think there are some under exploited (though less exciting) use cases, but all I see if big promises with under delivery. I would love to be proven wrong.
1. Content Generation:
LLMs can be used to generate high-quality, human-like content such as articles, blog posts, social media posts, and even short stories. Businesses can leverage this capability to save time and resources on content creation, and improve the consistency and quality of their online presence.
2. Customer Service and Support:
LLMs can be integrated into chatbots and virtual assistants to provide fast, accurate, and personalized responses to customer inquiries. This can help businesses improve their customer experience, reduce the workload on human customer service representatives, and provide 24/7 support.
3. Summarization and Insights:
LLMs can be used to analyze large volumes of text data, such as reports, research papers, or customer feedback, and generate concise summaries and insights. This can be valuable for businesses in fields like market research, financial analysis, or strategic planning.
4. HR Candidate Screening:
Use case: Using LLMs to assess job applicant resumes, cover letters, and interview responses to identify the most qualified candidates. Example: A large retailer integrating an LLM-based recruiting assistant to help sift through hundreds of applications for entry-level roles.
5. Legal Document Review:
Use case: Employing LLMs to rapidly scan through large volumes of legal contracts, case files, and regulatory documents to identify key terms, risks, and relevant information. Example: A corporate law firm deploying an LLM tool to streamline the due diligence process for mergers and acquisitions.
Spam isn't a feature. See also, this whole message that could just have been the headlines.
So… less clear than the website and not empowered to do anything (beyond ruining your reputation) because even you don't trust it?
See 1, spam isn't a feature. This is just trying to undo the damage from that (and failing).
If it's worth doing, it's worth doing well.
This seems unecessarily negative to me.
I'm working on AI tools for teachers and I can confidently say that GPT is just unbelievably good at generating explanations, exercises, quizes etc. The onus to review the output is on the teacher obviously, but given they're the subject matter experts, a review is quick and takes a fraction of the time that it would take to otherwise create this content from scratch.
It is negative. Because the rest of us are still forced to wade through the endless worthless sludge your ilk produces.
reducing the load on overworked teachers by using GPT to generate exercises, quizes and explanations for students is "endless worthless sludge"?
I have teachers in my family, their lives have been basically ruined by people using ChatGPT-4 to cheat on their assignments. They spend their weekend trying to workout if someone has "actually written" this or not.
So sorry, we're back to spam generator. Even if it's "good spam".
a bit dramatic. there has to be an adjustment of teaching/assessing, but nothing that would "ruin" anyone's life.
is it spam if it's useful and solves a problem? I don't agree it fits the definition any more.
Teachers are under immense pressure, GPT allows a teacher to generate extension questions for gifted students or differentiate for less capable students, all on the fly. It can create CBT material tailored to a class or even an individual student. It's an extremely useful tool for capable teachers.
If you don't have the power to just change your mind about what the entire curriculum and/or assessment context is, it can be a workload increase of dozens of hours per week or more. If you do have the power, and do want to change your entire curriculum, it's hundreds of hours one-time. "Lives basically ruined" is an exaggeration, but you're preposterously understating the negative impact.
Whether or not it's useful has nothing to do with whether or not it's spam. I'm not claiming that your product is spam -- I'll get back to that -- but your reply to the spam accusation is completely wrong.
As for your hypothesis, I've had interactions where it did a good job of generating alternative activities/exercises, and interactions where it strenuously and lengthily kept suggesting absolute garbage. There's already garbage on the internet, we don't need LLMs to generate more. But yes, I've had situations where I got a good suggestion or two or three, in a list of ten or twenty, and although that's kind of blech, it's still better than not having the good suggestions.
I think it has a lot to do with it. I can't see how generating educational content for the purpose of enhancing student outcomes with content reviewed by expert teachers can fall under the category of spam.
I like to present concrete examples of what I would consider to be useful content for a k-12 teacher.
Here's a very quick example that I whipped up
https://chatgpt.com/share/ec0927bc-0407-478b-b8e5-47aabb52d2...
This would align with Year 9 Maths for the Australian Curriculum.
This is an extremely valuable tool for
- A graduate teacher struggling to keep up with creating resources for new classes
- An experienced teacher moving to a new subject area or year level
Bear in mind that the GPT output is not necessarily intended to be used verbatim. A qualified specialist teacher with often times 6 years of study (4 year undergrad + 2 yr Masters) is the expert in the room who presumably will review the output, adjust, elaborate etc.
As a launching pad for tailored content for a gifted student, or lower level, differentiated content for a struggling student the GPT response is absolutely phenomenal. Unbelievably good.
I've used Maths as an example, however it's also very good at giving topic overviews across the Australian Curriculum.
Here's one for: elements of poetry:structure and forms
https://chatgpt.com/share/979a33e5-0d2d-4213-af14-408385ed39...
Again, an amazing introduction to the topic (I can't remember the exact curriculum outcome it's aligned to) which gives the teacher a structured intro which can then be spun off into exercises, activities or deep dives into the sub topics.
This is a result of poor prompting. I'm working with very structured, detailed curriculum documents and the output across subject areas is just unbelievably good.
This is all for a K-12 context.
There are countless existing, human-vetted, designed on special purpose, bodies of work full of material like the stuff your chatgpt just "created". Why not use those?
Also, each of your examples had at least one error, did you not see them?
I didn't could you point them out?
As a classroom teacher I can tell you that piecing together existing resources is hard work and sometimes impossible because resource A is in this text book (which might not be digital) and resource B is on that website and quiz C is on another site. Sometimes it's impossible or very difficult to put all these pieces together in a cohesive manner. GPT can do all that an more.
The point is not to replace all existing resources with GPT, this is all or nothing logic. It's another tool in the tool belt which can save time and provide new ways of doing things.
is it spam if it's useful and solves a problem? I don't agree it fits the definition any more.
Who said generating an essay is useful sorry ? What problem does that solve?
Your comments come accross as overly optimistic and dismissive . Like you have something to gain personally and aren’t interested in listening to others feedback.
I'm developing tools to help teachers generate learning material, exercises and quizes tailored to student needs.
Useful learning materials aligned with curriculum outcomes, taking into account learner needs and current level of understanding is literally the bread and butter of teaching.
I think those kinds of resources are both useful and solve a very real problem.
Fair point. I do have something to gain here. I've given a number of example prompts that are extremely useful for a working teacher in my replies to this thread. I don't think I'm being overly optimistic here. I'm not talking vague hypotheticals here, the tools that I'm building are already showing great usefulness.
One potential fix, or at least a partial mitigation, could be to weight homework 50% and exams 50%, and if a student's exam grades differ from their homework grades by a significant amount (e.g. 2 standard deviations) then the lower grade gets 100% weight. It's a crude instrument, but it might do the job.
Why haven’t they just gone back to basics and force students to write out long essays on paper by hand and in class?
Also have teachers in my family. Most of the time is spent adjusting the syllabus schedule and guiding (orally) the stragglers. Exercises, quizes and explanations are routine enough that good teachers I know can generate them on the spot.
Every year there are thousands of graduate teacher looking for tools to help them teach better.
Even the best teacher can't create an interactive multiple choice quiz with automatic marking, tailored to a specific class (or even a specific student) on the spot.
I've been teaching for 20+ years, I have a solid grasp of the pain points.
Neither can "AI" though, so what's the point here?
I'm creating tools on top of AI that can which is my point.
Can you post a question and answer example if it doesn’t violate NDA because I have very little faith this is good for students.
sure
here's an example of a question and explanation which aligns to Australian Curriculum elaboration AC9M9A01_E4 explaining why frac{3^4}{3^4}=1, and 3^{4-4}=3^0
https://chatgpt.com/share/89c26d4f-2d8f-4043-acd7-f1c2be48c2...
to further elaborate why 3^0=1 https://chatgpt.com/share/9ca34c7f-49df-40ba-a9ef-cd21286392...
This is a relatively high level explanation. With proper prompting (which, sorry I don't have on hand right now) the explanation can be tailored to the target year level (Year 9 in this case) with exercises, additional examples and a quiz to test knowledge.
This is just the first example I have on hand and is just barely scratching the surface of what can be done.
The tools I'm building are aligned to the Austrlian Curriculum and as someone with a lot of classroom experience I can tell you that this kind of tailored content, explanations, exercises etc are a literal godsend for teachers regardless of experience level.
Bear in mind that the teacher with a 4 year undergrad in their specialist area and a Masters in teaching can use these initial explanations as a launching pad for generating tailored content for their class and even tailored content for individual students (either higher or lower level depending on student needs). The reason I mention this is because there is a lot of hand-wringing about hallucinations. To which my response is:
- After spending a lot of effort vetting the correctness of responses for a K-12 context hallucinations are not an issue. The training corpus is so saturated with correct data that this is not an issue in practice.
- In the unlikely scenario of hallucination, the response is vetted by a trained teacher who can quickly edit and adjust responses to suit their needs
Let’s call it for what it is- taking poorly organized existing information and making it organized and interactive.
“Here are some sharepoint locations, site Maps, and wikis. Now regurgitate this info to me as if you are a friendly call center agent.”
Pretty cool but not much more than pushing existing data around. True AI I think is being able to learn some baseline of skills and then through experience and feedback adapt and be able to formulate new thoughts that eventually become part of the learned information. That is what humans excel at and so far something LLMs can’t do. Given the inherent difficulty of the task I think we aren’t much closer to that than before as the problems seem algorithmic and not merely hardware constrained.
Which is extremely valuable!
Don't underestimate how valuable it is for teachers to do exactly that. Taking existing information, making it digestable, presenting it in new and interseting ways is a teacher's bread and butter.
It’s valuable for use cases where the problem is “I don’t know the answer to this question and don’t know where to find it.” That’s not in and of itself a multibillion dollar business when the alternative doesn’t cost that much in the grand scheme of things (asking someone for help or looking for the answer).
Are you suggesting a chatbot is a suitable replacement for a teacher?
As a teacher - I have no shortage of exercises, quizes etc. Internet is full of this kind of stuff and I have no trouble finding more than I ever need. 95% of my time an mental capacity in this situation goes for deciding what makes sense in my particular pedagogical context? What wording works best for my particular students? Explanations are even harder. I find out almost daily that explanations which worked fine in last year, don't work any more and I have to find a new way, because previous knowledge, words they use and know etc of new students are different again.
Which all takes valuable time us teachers are extremely short on.
I've been a classroom teacher for more than 20 years, I know how painful it is to piece together a hodge podge of resourecs to put together lessons. Yes the information is out there, but a one click option to gather this into a cohesive unit for me saves me valuable time.
Which is exactly what GPT is amazing at.Brainstorming, rewriting, suggesting new angles of approach is GPTs main stength!
Prompting GPT to give useful answers is part of the art of using these new tools. Ask GPT to speak in a different voice, take on a persona or target a differnt age group and you'll be amazed at what it can output.
Exactly! Reframing your own point of view is hard work, GPT can be an invaluable assistant in this area.
I’ve rarely if ever seen a model fully explain mathematical answers outside of simple geometry and algebra to what I would call an adequate level. It gets the answer right more often than explaining why that is the correct answer. For example, it finds a minimal case to optimization, but can’t explain why that is the minimal result among all possibilities.
Dear lord, if someone started relying on LLMs for legal documents, their clients would be royally screwed…
They're currently already relying on overworked, underpaid interns who draft those documents. The lawyer is checking it anyway. Now the lawyer and his intern have time to check it.
I have no idea what type of law you're talking about here, but (given the context of the thread) I can guarantee you major firms working on M&As are most definitely not using underpaid interns to draft those documents. They are overpaid qualified solicitors.
Apologies I mean candidate attorneys when I say interns. Those overpaid qualified attorneys, read it and sign off on it.
I suggest we do not repeat the myth and urban legend that LLMs are good for legal document review. I had a couple of real use cases used for real clients who were hyped about LLMs to be used for document review and trying to save salary, for Engish language documents. We've found Kira, Luminance and similar due diligence project management stuff as useful being a timesaver if done right. But not LLMs. Due to longer context windows, it is possible to ask LLMs the usual hazy questions that people ask in a due diligence review (many of which can be answered dozens of different ways by human lawyers). Is there a most favoured nation provision in the contract, is there a financial cap limiting the liability of the seller or the buyer, governing law etc. Considering risks of uploading such documents into ChatGPT, you are stuck with Copilot M365 etc. or some outrageously expensive "legal specific" LLMs that I cannot test. Just to be curious with Copilot I've asked five rather simple questions for three different agreements (where we had the golden answer), and the results were quite unequal, but mostly useless - in one contract, it incorrectly reported for all questions that these cannot be answered based on the contract (while the answers were clearly included in the document), in an another, two questions were answered correctly, two questions not answered precisely (just governing law being US instead of the correct answer being Michigan, even after reprompting to give the state level answer, not "USA") and hallucinated one answer incorrectly. In the third one, three answeres were hallucinated incorrectly, answered one correctly and one provision was not found. Of course, it's better to have a LEGAL specific benchmark for this, but 75% hallucination in complex questions is not something that helps your workflow (https://hai.stanford.edu/news/hallucinating-law-legal-mistak...) I don't recommend at least LLMs to anyone for legal document reviews, even for the English language.
I'm not talking about reviewing, only drafting. Every word should be checked. A terrible idea relying on the advice of an LLM.
Except for number 3, the rest are more often disastrous or insulting to users and those depending on the end products/services of these things. Your reasoning is so bad that i'm almost tempted to think you're spooning out PR-babble astro-turf for some part of the industry. Here's a quick breakdown:
1. content: Nope, except for barrel-bottom content sludge of the kind formerly done by third world spam spinning companies, most decent content creation stays well away from AI except for generating basic content layout templates. I work as a writer and even now, most companies stay well away from using GPT et al for anything they want to be respected as content. Please..
2. Customer service: You've just written a string of PR corporate-speak AI seller bullshit that barely corresponds to reality. People WANT to speak to humans, and except for very basic inquiries, they feel insulted if they're forced into interaction with some idiotic stochastic parrot of an AI for any serious customer support problems. Just imagine some guy trying to handle a major problem with his family's insurance claim or urgently access money that's been frozen in his bank account, and then forced to do these things via the half-baked bullshit funnel that is an AI. If you run a company that forces that upon me for anything serious in customer service, I would get you the fuck out of my life and recommend any friend willing to listen does the same.
3. This is the one area where I'd grant LLMs some major forward space, but even then with a very keen eye to reviewing anything they output for "hallucinations" and outright errors unless you flat out don't care about data or concept accuracy.
4. For reasons related to the above (especially #2) what a categorically terrible, rigid way to screen human beings with possible human qualities that aren't easily visible when examined by some piece of machine learning and its checkbox criteria.
5. Just, Fuck No... I'd run as fast and far as possible from anyone using LLMs to deal with complex legal issues that could involve my eventual imprisonment or lawsuit-induced bankruptcy.
2.I think you overestimate the caliber of query received in most call centres. Even when it comes to private banks (for those who've been successful in life), the query is most often something small like holding their hand and telling them to press the "login" button.
Also these all tend to have an option where you simply ask it and it will redirect you to a person.
Those agents deal with the same queries all day, despite what you think your problem likely isn't special, in most cases may as well start calling the agents "stochastic parrots" too while you're at it.
I’ve been doing RLHF and adjacent work for 6 months. The model responses across a wide array of subject matter are surface level. Logical reasoning, mathematics, step by step, summarization, extraction, generation. It’s the kind of output the average C student is doing.
We specifically don’t do programming prompts/responses nor advanced college to PHD level stuff, but it’s really mediocre at this level and these subject areas. Programming might be another story, I can’t speak to that.
All I can go off is my experience but it’s not been great. I’m willing to be wrong.
Is the output of average C students not commercially valuable in the listed fields? If AI is competing reliably with students then we've already hit AGI.
IMO the unreasonable uselessness of LLMs is because for most tasks involving language the accuracy needs to be unbelievably high to have any real value at all.
We just don't have that.
We have autocomplete on steroids and many people are fooling themselves that if you just take more steroids you will get better and better results. The metaphor is perfect because if you take more and more steroids you get less and less results.
It is why in reality we have had almost no progress since April 2023 and chatGPT 4.
Calling LLMs autocompleters is an insult to autocompleters.
I don't find coding assistants to be very useful. Image generation was fun for a few weeks. Speech recognition is useful.
Anyway, considering all these things can be done on device, where is the long term business prospect of which you speak?
I've come to notice a correlation between contemporary AI optimism and having effectively made the jump to coding with AI assistants.
I think this depend heavily on what type of coding your doing. The more your job could be replaced by copy/pasting from Stack Overflow, the more useful you find coding assistants.
For that past few years most of the code I've written has been solving fairly niche quantitative problems with novel approaches and I've found AI coding assistants to range from useless to harmful.
But on a recent webdev project, they were much more useful. The vast majority of problems in webdev are fundamentally not unique so a searchable pattern library (which is what an LLM coding assistant basically is) should be pretty effective.
For other areas of software, they're not nearly as useful.
I think this is true and also why you see some "older devs just don't like AI" type comments. AI assistants seem to be great at simple webdev tasks, which also happens to be the type of work that more junior developers do day to day.
I have also found them useful with that and I keep one active for those types of projects because of the speed up, although I still have to keep a close eye on what it wants to inject. They also seem to excel at generating tests if you have already developed the functions.
Then there are more difficult (usually not webdev) projects. In those cases, it really only shines if I need to ask it a question that I would previously have searched on SO or some obscure board for an answer. And even then, it really has to be scrutinized, because if it was simple, I wouldn't be asking the question.
There is def. something there for specific types of development, but it has not "changed my life" or anything like that. It might have if I was just starting out or if I only did webdev type projects.
As an "older dev" who doesn't like AI, the thing that annoys me most is the UX is horrible. It's like arguing in chat with an extremely overconfident junior dev who isn't capable of learning or improving with time and experience. That's just a miserable way to spend time. I'd rather spend that time thinking clearly about the problem, and then writing down the solution (clearly).
If this thing also conferred an actual productivity advantage that would be one thing, and it might motivate me to get past the horrible UX, but I haven't seen any evidence yet.
I fear the approach that maximises productivity is a literal one-shot approach: Give the LLM one or two shots at generating a somewhat passable first attempt (including all or at least most of the boilerplate) and then strictly fix up stuff yourself. I recently spend a day attempting to build a relatively simple GUI for a project which _maybe_ contains a couple of days of programming work. It got the gist of the GUI basically in one. And the next two or three prompts then added the buttons I wanted. Most of it even worked
But after that we ran into a kind of loop, where you put my feelings into much better words than I could. If I had stopped after iteration 3, I probably would have finished what I wanted to do in half a day
Now try to mute a video on youtube and understand what's being said from the automatic subtitles.
If you do it in english, be aware that it's the best performing language and all others are even worse.
For some reason, YouTube is not using a very good STT system now. The lack of sentence punctuation is particularly annoying. Transcriptions by Whisper and Gemini 1.5 Pro are much better. From a couple of weeks ago:
https://news.ycombinator.com/item?id=41199567#41201773
I expect that YouTube will up their transcription game soon, too.
I've tried whisper too. I made this: https://codeberg.org/ltworf/srtgen
Basically it's kinda useful to put time tags, but I need to manually fix each and every sentence. Sometimes I need to fix the time tags as well.
I just spoke about youtube because it's more popular and easy to test.
Sometimes speech-to-text machine learning models give very good results, however I think the key is that:
1. It's overwhelmingly more useful than the [no text] it was replacing, particularly for the deaf or if you want to search for keywords in a video.
2. When it fails, it tends to do so in ways that trigger human suspicion and oversight.
Those aren't necessarily true of some of the things people are shoehorning LLMs into these days, which is why I'm a lost more pessimistic about that technology.
Just today, I received a note from a gas technician, part handwritten, for the life of me I couldn't make out what he was saying, I asked ChatGPT and it surprisingly understood, rereading the original note I'm very sure it was correct.
“This time is different” in one fundamental, methodological, epistemological way: we test on the training set now.
This has follow-on consequences for a shattering phase transition between “persuasive demo” and “useful product”.
We can now make arbitrarily convincing demos that will crash airplanes (“with no survivors!”) on the first try in production.
This is institutionalized by the market capitalizations of 7 companies being so inflated that if they were priced accurately the US economy would collapse.
There was really one AI winter which is a sample size of 1, saying this time is different is justified based on the exponential improvement over AI 10 years back
See this is exactly what is wrong with “this time it’s different” here. AI has been useful and used for decades (but under a different name because the term was tainted by previous bubbles). Look at the section “AI behind the scenes” here https://en.wikipedia.org/wiki/History_of_artificial_intellig...
There were products and "path to products" too. Once the hype died down nobody wanted them. It is the same this time.
"AGI" is a nonsense term anyway. Humans don't have "general" intelligence either: our intelligence is specialized to our environment.
Humans is the most general intelligence we know about, so that is why we called it general intelligence, because we have made so many intelligences that are specialized on a specific domain like calculators or chess engines we need a word for something that is as general as humans, because being able to replace humans is a very important goal.
Yes, humans are the most general intelligence we know about. That doesn't say much about how general it is, just highlights our limitations.
This is a bit like saying Earth isn't big, because there are far larger planets etc. out there. For the average conversation, Earth is "big".
"AGI" means many different things to many different people: to me any AI which is general is an AGI so GPT-3.5 counts; to OpenAI it has to be economically transformative to count; to some commentators here it has to be superhuman to count.
I think that none of the initials are boolean; things can be degrees of artificial, degrees of general, and degrees of intelligent.
I think most would assert that humans count as a "general" intelligence, even if they disagree about most of the other things I've put in this comment.
I have been using Chat-GPT has a full time expert and I can unequivocally tell you that its a transformative piece of technology. The technology isn't hyped.
I agree as this is also my personal experience. But I also see the usage of ChatGPT is falling down fast from 1.8 billion visitors to 260 million last month [1].
I am probably through some ETF an investor in MS, so I do hope the openai API usage is showing a more stable and upward trend.
[1]: https://explodingtopics.com/blog/chatgpt-users
Well ChatGPT is no longer the top dog and there's quite a bit of competition in the space. Including Llama 3.1 which is free. In general I think most of the moat that OpenAI had has evaporated in the last few months, but also for other LLM companies.
Not sure how they plan on making money in the long-term, eventually the investors and shareholders will start asking when they will be seeing the returns on their investment.
It is very nice as "Markov chains on steroids", but people believing that LLMs are anything but a distracting local maximum on the path to AGI are 200% in kool-aid drinking mode.
All Kahneman's system 2 thinking is just slow deliberate thinking. And these models do indeed have this characteristic to an extent, as evidenced with chain of thought reasoning.
You can see this in action with multiplication. Much like humans when asked to guess the answer, they'll get it wrong, unless they know the answer from rote learning multiplication tables, this System-1 thinking. In many cases when asked they can reason further and solve it, by breaking it down and solving it step by step, much like a human, this is system-2 thinking.
In my opinion, it seems nearly everything is there for it it to take the next leap in intelligence, it's just putting it all together.
Agreed. System 2 strategizing may simply be the recursive application of symbolic System 1 tooling. An LLM is entirely capable of reading a problem, determining the immediate facts and tokens (fast and intuitive system 1) and determining the ideal algorithm to resolve them (logical analytical system 2). The execution to do all those steps at once is lacking in current LLMs (debatably - they get better every month) - but any basic architecture breaking things down into component sub-questions clearly works.
If you look at AI history there is often fairly steady progress in a given skill area for example chess programs improved in a steady way on ELO scores and you could project pretty well the future by drawing a line on a graph. Similarly large language models seem to be progressing from toddler like to high school student like (now) to PhD like - shortly. There are skills AI are still fairly bad at like the higher level reasoning you mention, and in robot form being able to pop to the shops to get some groceries say but I get the impression those are also improving in a steady way and it won't be so long.
For readers' edification, would you mind making a strong hypothetical argument for why this time it actually is different, from an expert's perspective?
I am yet to see any reasoned argument for why it is easy to build real AI and that it will come fast.
As you said, AI has been there for decades and stagnated for pretty much the whole time. We've just had a big leap, but nothing says (except BS hype) that we're not in for a long plateau again.
We have "real ai" already.
As for future progress, have you tried just simple interpolation of the progress so far? Human level intelligence is very near. (Though of course artificial intelligence will never exactly match human intelligence: it will be ahead/behind in certain aspects...)
- We don't have a "real AI" at all. Where's Skynet, where's HAL-9000? Where are the cute robotic butlers from the "I, Robot" movie?
- Simple interpolation of the progress is exactly the problem here. Look at the historical graphs of AI funding and tell me with a straight face that we absolutely must use simple interpolation.
- Nope, human-level intelligence is not even close. It remains as nebulous and out of reach as ever. ChatGPT's imitation of intelligent speech falls apart very quickly when you chat with it for more than a few questions.
You shouldn’t use science fiction as your reference point. It’s like saying “where is my flying car?” (Helicopters exist)
Why shouldn't I? Humans developed this intelligence naturally (as far as we know). Many people claim we're intelligent enough to repeat the process with artificial organisms, and guide it, and perfect it. I want to see it.
And btw in the Terminator novelizations it was clearly stated that Skynet was a very good optimization machine but lacked creativity. So it's actually a good benchmark: can we create an intelligent machine that needs no supervision but still has limitations (i.e. it cannot dramatically reformulate its strategy in case it cannot win, which is exactly what happened in the books)?
Just because someone can tell a convincing story, doesn’t mean reality (once technology catches up) will resemble the devices of that story. Science fiction is fiction, and unconstrained by the annoying restrictions of physical reality.
That’s the point of my flying car comparison. We HAVE flying cars: they’re called helicopters. Because as it turns out there is just no physical way to make a vehicle in the form factor of a car fly, except by rotary wing. But people will still say “where’s my flying car?” because they are hung up on reality resembling science fantasy, as you are.
We have AI. We even already have AGI. It just doesn’t resemble the Terminator, because The Terminator is a made up story disconnected from reality.
And this is why, I feel, I can never discuss with the AI fans. They are happy to invent their own fiction while berating popular fiction in the same breath.
No, we really don't have AGI. Feel free to point out some of humanity's pressing problems being trivially solved today with it, please. I'll start: elderly people care, and fully automated logistics.
I’m not an “AI fan.” But anyway.
Artificial. General. Intelligence.
The term, as originally defined, is for programs which are man-made (Artificial), able to efficiently solve problems (Intelligence), including novel problem domains outside those considered in its creation (General). Artificial General Intelligence, or AGI. That’s literally all AGI means, and ChatGPT absolutely fits the bill.
What you describe is ASI, or artificial super intelligence. In the late 90’s, 00’s, and early 10’s, a weird subgroup of AI nerds got it into their head that merely making an AGI (even a primitive one) would cause a self-recursion improvement loop and create ASI in short order. They then started saying “achieve AGI” as a stand in for “emergence of ASI” as the two were intricately linked in their mind.
In reality the whole notion of AGI->ASI auto-FOOM has been experimentally discredited, but the confusion over terminology remains.
Furthermore, the very idea of ASI can’t be taken for granted. A machine that trivially solves humanity’s pressing problems makes nice sci-fi, but there is absolutely no evidence to presume such a machine could actually exist.
You are addressing the wrong person if you think I give two hoots how many more acronyms will the AI area invent to conceal the fact that the only thing they actually achieved is remove a lot of artists from the job market.
I don't care how it's called. We don't have it. I am not "confused over terminology", I want to see results and yet again they don't exist. Let's focus on results.
Sure. Because we actually have this super-intelligence already and we can compare with it, right? Oh wait, no we don't. So what's your point? Some people gave up and proclaimed that it can't be done? Like we haven't seen historical examples of this meaning exactly nothing, hundreds of times already.
Look, we'll never be able to talk about it before you stop confusing industry gate-keepers who learned how to talk to get VC money and obfuscate reality with, you know, the actual reality in front of us. You got duped by the investor talk and by the scientists never wanting to admit their funding might have been misplaced by being given to them, I am afraid.
Finally, nope, again and again, we don't have AGI even if I accept your definition. Show me a bot that can play chess, play StarCraft 2, organize an Amazon warehouse item movements and shipping, and coordinate a flight's touch-down with the same algorithms / core / whatever-you-want-to-call it. Same one, not different ones. One and the same.
No? No AGI then either.
The people in the bronze age could have easily said "there is no evidence we would be able to haul goods while only pressing pedals and rotating a wheel". That's not an argument for anything at all, it's a short-sighted assertion that we might never progress that's only taking the present and the very near future into account. Well, cool, you don't believe it will happen. And? That's not an interesting thing to say.
Other people didn't believe we could go to the Moon. We still did. I wonder how quickly the naysayers hid under the bed after that so nobody could confront them about it. :D
But anyway. I got nothing more to say to people who believe VC talk and are hell-bent on inventing many acronyms to make sure they are never held accountable.
I for one want machines that solve humanity's problems. I know they can exist. I know nearly nobody wants to work on them because everybody is focused on the next quarter's results. All this is visible and well-understood yet people like you seem to think that this super narrow view is the best humanity can achieve.
Well, maybe it's the best you can achieve. I know people who can do more.
You are being unnecessarily aggressive. I won’t be continuing this debate.
That is likely true. We can't agree on basic premises so there's no point pursuing a discussion regardless of the tone.
Inventing a fricken time machine wasn't creative?
I know right? :D That always bothered me as well.
In the novelizations it was written that Skynet could not adapt to humans not running away or not vacating territories after they have been defeated. One of the quotes was: "Apparently it underestimated something that it kept analyzing even now: the human willpower." I've read this as Skynet not being able to adapt against guerilla warfare -- the hit-and-run/hide tactics.
But the TL;DR was that Skynet was basically playing something like StarCraft as if it played against another bot, and ultimately lost because it played against humans. That was the "Skynet was not creative" angle in the novelizations.
This is a complete tangent but:
In Terminator 1 Skynet looses because John Conner taught people how to fight the machines, but John Conner only knows this because Kyle Reese taught Sarah Conner how to fight the machines and she taught John Conner. But Kyle Reese only knows this because he was taught by John Conner- so there's no actual source of the information on how to fight the machines, it's a loop with no beginning or end.
I had a philosophy teacher who said this is evidence of divine intervention to destroy Skynet, essentially God told people through John Conner how to win, but a cut scene in Terminator 1 implies Skynet was also created by reverse engineering the chip in the destroyed Terminator- implying there's also no origin of the information on how to create Skynet and it's also an infinite loop.
Yeah, these discussions are fascinating but I'd still think it's not very hard to learn how to blow stuff up and sabotage assembly lines, given enough tries.
So it's not exactly an infinite loop IMO, it's more like that the first iteration was more crude and the machines were difficult to kill but then people learned and passed the information along back in time, eventually forming the infinite loop -- it still had a first step though, it didn't come out of nothing.
To be fair, I’ve talked to a lot of people who cannot consistently perform at the mistral-12b level.
I think we expect AGI to be much smarter than the average joe, and free of occasional stupidity.
What we’ve got is an 85IQ generalist with unreliable savant capabilities, that can also talk to a million people at the same time without getting distracted. I don’t see how that isn’t absolutely a fundamental shift in capability.
It’s just that we expect it to be spectacularly useful. Not like homeless joe, who lives down by the river. Unfortunately, nobody wants a 40 acre call center of homeless joes, but it’s hard to argue that HJ isn’t an intelligent entity.
Obviously LLMs don’t yet have a control and supervision loop that gives them goal directed behaviour, but they also don’t have a drinking problem and debilitating PTSD with a little TBI thrown in from the last war.
It’s not that we aren’t on the cusp of general intelligence, it’s that we have a distorted idea of how useful that should be.
Very shallow assessment, first of all it's not a generalist at all, it has zero concept of what it's talking about, secondly it gets confused easily unless you order it to keep context in memory, and thirdly it can't perform if it does not regularly swallow petabytes of human text.
I get your optimism but it's uninformed.
I can find you an old-school bot that performs better than uneducated members of marginalized and super poor communities, what is your example even supposed to prove?
What's HJ? If it's not a human then it's extremely easy to argue that it's not an intelligent entity. We don't have intelligent machine entities, we have stochastic parrots and it's weird to pretend otherwise when the algorithms are well-known and it's very visible there's no self-optimization in there, there's no actual learning, there's only adjusting weights (and this is not what our actual neurons do btw), there's no motivation or self-drive to continue learning, there's barely anything that has been "taught" to combine segments of human speech and somehow that's a huge achievement. Sure.
Nah, we are on no cusp of general AGI at all. We're not even at 1%. Don't know about you but I have a very clear idea what would AGI look like and LLMs are nowhere near. Not even in the same ballpark.
It helps that I am not in the area and I don't feel the need to pat myself on the back that I have managed to achieve the next AI plateau which the area will not soon recover from.
Bookmark this comment and tell me I am wrong in 10 years, I dare you.
What is intelligence? We must have very different definitions!
Nobody knows what intelligence actually is. But asking this philosophical question and your discussion opponent not having a clear answer is a very obvious discussion trap and a discussion shut-down and it does NOT immediately follow that your claim -- "we have AI / AGI" -- becomes automatically true. It does not.
And I am pretty sure my own intelligence goes much farther than regurgitating text that I have no clue about (like ChatGPT does not have symbol logic that links words with objects it can "feel" physically or otherwise).
HJ is Homeless Joe, an inference that a 12b stochastic text generator would not have missed lol. But sure, ill reflect in 10 years.
TBH I hope im wrong, and that there is magic in HJ that makes him special in the universe in a way that GPT26 can never be. But increasingly, I doubt this premise. Not because of the "amazing capabilities of LLMs" which i think are frequently overstated and largely misunderstood, but more because of the dumbfounding shortcomings of intelligent creatures. We keep moving the bar for AGI, and now AGI is assumed to be what any rational accounting would classfy as ASI.
Where we are really going to see the bloom of AI is in goal directed systems, and I think those will come naturally with robotics. I predict we are in for a very abrupt 2nd industrial revolution, and you and I will be able to have this discussion either over a 55 gallon barrel of burning trash, or in our robot manicured botanical gardens sometime in the near future lol.
good times, maybe. Interesting times , for sure.
We have found common ground.
Yes, a lot of us utilize defective judgments, myself included, fairly often. My point was that LLMs, for all their praise, can't even reach 10% of an average semi-intelligent organic being.
I don't know who is "we" (and I wish people stopped pretending that "we" are all a homogenous mass) but I've known what an AGI should be ever since I've watched movies about Skynet and HAL-9000. ¯\_(ツ)_/¯
Secondly, it's the so-called "AI practitioners" who constantly move the goal posts (now there's "ASI"? -- you know what, I actually don't want to know) because they're periodically being called out and can't hide the fact that they have nearly nothing again. So what's better than obfuscating that fact by having 100+ acronyms? It's a nice cover and apparently there are still investors who are buying it. I get it, we have to learn to say the right things to get funding.
I agree. Physical feedback is needed if we want an electronic entity to "evolve" similarly to us.
I agree this is 100% inevitable but I don't think it's coming as soon as you say. The LLMs are hopelessly stuck even today and the whole AI area will suffer for it for a while after the bubble bursts... which is the event that I am certain is coming soon.
I don't think they'll use LLM's for customer service.
But it's a building block. And when used well it may be possible to get to zero hallucinations and good accuracy in question answering for limited domains - like the call center.
If the current LLMs manage to achieve even only that it would be an enormous win. Alas they still have not.
This is honestly one of the most gpt-2 things I’ve ever read.
Well he gave you a list of credentials of why you should believe him. Isn’t that enough ?
This has to do with ML and numerical computing, how?
Well I was being sarcastic. I really dislike it when people have to first convince you that you should trust them.
Either make a good argument or don’t.
Argument from authority is a pernicious fallacy, and typically effective too. You were right to call it out. I must admit I overlooked the sarcasm, however.
Don't feel too bad; until I read the next response I was in two minds about whether sarcasm was intended or not.
It's bloody hard to tell, sometimes :-/
it can be for some
https://en.m.wikipedia.org/wiki/Poe%27s_law
Human beings can't evaulate the truth of things based only on the argument. Persuasive liars, cons, and incompetents are a very known phenomenon. Most of human history we misunderstood nature and many other things because we relied on 'good arguments'. Not that we need it, but research shows that human intuition about the truth of something isn't good without expertise.
When I need medical advice, I get it from someone who has convinced me that they have expertise; I don't look for 'good arguments' that persuade me, because I don't know what I'm talking about.
I have expertise in other things. In those fields, I could easily persuade people without it of just about anything. (I don't; I'm not a sociopath.) I imagine anyone with professional expertise who reads this can do the same.
Oh, I was bamboozled.
If people say "believe me" at the end of every second sentence you should doubt them. Not thinking of anyone in particular.
No. Stating credentials proves nothing at all. Even less so on the internet.
edit: oh sorry I didn't get that it was sarcasm
Let’s poll RLHF workers since they actually see the tech the most.
The thing that's different this time is the hardware capacity in TFLOPs and the like passing human brain equivalence.
There's a massive difference between much worse than human AI - a bit meh, and better than human AI - changes everything.
It probably won't be easy but the huge value of better than human AI will ensure loads of the best and brightest working on it.
You sound like you don't actually understand anything about LLMs and are buying into the hype. They are not cognizant let alone conscious. They don't understand anything. The tokens could be patterns of colored shapes with no actual meaning, only statistical distributions and nothing about how the LLMs work would change.
I can put your brain in a vat and stimulate your sensory neurons with a statistical distribution with no actual meaning, and nothing about how your brain works would change either.
The LLM and your brain would attempt to interpret meaning with referent from training, and both would be confused at the information-free stimuli. Because during "training" in both cases, the stimuli received from the environment is structured and meaningful.
So what's your point?
By the way, pretty sure a neuroscientist with 20 years of ML experience has a deeper understanding of what "meaning" is than you do. Not to mention, your response reveals a significant ignorance of unresolved philosophical problems (hard problem of consciousness, what even is meaning) which you then use to incorrectly assume a foregone conclusion that whatever consciousness/meaning/reasoning is, LLMs must not have it.
I'm partial to doubting LLMs as they are now have the magic sauce, but it's more that we don't actually know enough to say otherwise, so why state that we do know?
We can't even say we know our own brains.
Your response is nonsense. We don't know how consciousness arises from matter but we do have significant understandings about knowledge, reasoning, modeling, visualization, object permanence, etc. etc. that are significant parts of how the human mind works. And we know LLMs have none of these.
The point of my colored shape example is that it is an illusion that there is anything resembling a mind inside an LLM. I thought that was obvious enough I didn't need to explain it further than I did.
As far as the original commenter's credentials; there's lots of people who should know better but buy into hype and nonsense.
Go ahead and cite your sources. For every study claiming that LLMs lack these qualities, there are others that support and reinforce the connectionist model of how knowledge is encoded, and with other parallels to the human brain. So... it's inconclusive. It's bizarre why you so strongly insist otherwise when it's clear you are not informed.
And my example with subjecting a human brain through your procedure is to illustrate what a garbage experiment design it is. You wouldn't be able to tell there's a mind inside either. Both LLM and human brain mind would be confused. Both would "continue working" in the same way, trying to interpret meaning from meaningless stimulation.
So you don't have a point to make, got it.
We know LLMs don't have those things prima facia because they fail at all of them constantly. We also know how they work, they are token predicters. That is all they are and all they can do. What can be accomplished with that is pretty cool, but humans love to imagine there is more going on when there isn't. Just like with Eliza.
If you don't understand how your attempt to apply my colored shape analogy to a human brain is nonsensical I am not going to waste my time explaining it to you. I had a point, I made it, and apparently it escaped you.
And if you don't see how the example with the human brain throws a wrench in your analogy, it would explain why you'd think it as nonsensical, as it's exactly of relevance.
Ah there it is, you've betrayed a deep lack of understanding in all relevant disciplines (neuroscience, cognition, information theory) required to even appreciate the many errors you've made here.
You sure understand the subject matter and have nothing possibly to learn. Enjoy.
https://en.wikipedia.org/wiki/Predictive_coding
I'm aware of the theories that LLM maximalists love to point to over and over that tries to make it seem like LLMs are more like human brains than they are. These theories are interesting in the context of actual minds but you far over extend their usefulness and application by trying to apply them to LLMs.
We know as a hard fact that LLMs do not understand anything. They have no capacity to "understand". The constant, intractible failure modes that they continuously exhibit are clear byproducts of this fact. By continuing to cling to the absurd idea that there is more going on than token prediction you make yourself look like the people who kept insisting there was more going on with past generation chat bots even after being shown the source code.
I have understood all along why you attempt to extend my colored shape example to the brain, but your basis for this is complete nonsense. Because a) we do not have the actual understanding of the brain to do this and b) it's competely beside the point, becuase we know that minds do arise from the brain. My whole point is an LLM is an illusion of a mind which is effective because it outputs words, which we are so hard wired to associate with other minds, expecially when they seem to "make sense" to us. If instead of words you use something nonsensical like colored shapes with no underlying meaning, this illusion of the mind goes away and you can see an LLM for whst it is.
What an absurd response. Yes, you'd probably cause the human brain to start malfunctioning terribly at the form of consciousness it's well -accustomed to managing within the context of its normal physical substrate and environment. You'd be doing that (and thus degenerating it badly) because you removed it from that ancient context whose workings we still don't well understand.
Your LLM on the other hand, has no context in which it shows such a level of cognitive capacity, higher-order reasoning, self direction and self awareness that we daily see humans to be capable of.
really? An appeal to authority? Many smart, educated people can still fall for utter nonsense and emotional attachment to bad ideas.
Yes, it will be confused as well, and for all outwards observable signs will fail to make sense of the stimuli, yet it will "aware" of its inability to understand, much like a human brain would.
If you doubt that, open a new session and type some random tokens, you will get the answer that it's confused.
Any other statement as to "consciousness" verges into the philosophical and unanswerable via empirical means.
And ah, to frame it as an appeal to authority when the topic is precisely the subject of a neuroscientist's study.
Sounds like you know a thing or two about nonsense and emotional attachment to bad ideas.
You persist in talking nonsense.
There is no empirical evidence of any awareness whatsoever in any LLM, at all. Even their most immersed creators don't make such a claim. An LLM itself saying anything about awareness doesn't mean a thing. It's literally designed to mimic in such a way. And you speak of discussions of consciousness being about the philosophical and unanswerable?
At least when talking about human awareness, one applies these ideas to minds that we personally as humans perceive to be aware and self-directed from our own experience (flawed as it is). You're applying the same notion to something that shows no evidence of awareness while then criticizing assumptions of consciousness in a human brain?
Such a sloppy argument indeed does make appeals to authority necessary I suppose.
It seems you've lost the train of your own argument.
You claim LLMs have no context at all in which it shows a similar level of cognitive capacity.
Yet clearly this claim is in contention with the fact that an LLM will indeed be able to evince this, much like a human brain would: by attesting to its own confusion. That is ostensibly empirical and evidential to a nonzero degree.
Thus your claim is too strong and therefore quite simply wrong. Claim mimicry? Then prove human brain consciousness does not derive from the process of mimicry in any form. You can't. In fact, the free energy principle, neuroscience's leading theory of human consciousness, argues the opposite: that prediction and mimicry encompass the entirety of what brains actually do. https://en.wikipedia.org/wiki/Predictive_coding
And no such claim was made--"awareness" was quoted for a reason.
Yes, as this was parent's claim: "They are not cognizant let alone conscious'.
And talking about sloppy argument--it may well turn out that something can be "designed to mimic" yet still be conscious. I'll leave that for you to puzzle out how on earth that might be possible. The exercise might help you form less sloppy arguments in the future.
No. But I suppose you've lost the plot a few inferential steps prior this so your confusion is not surprising.
Protip, instead of claiming everything that goes against your sensibilities as nonsense, perhaps entertain the possibility that you might just not be as well informed as you thought.
When Weizenbaum demonstrated Eliza to his colleagues, some thought there was an intelligent consciousness at the heart of it. Few even continued to believe this after they were shown the source code, which they were able to read and understand. Human consciousness is full of biases and the most advanced AI cannot reliably determine which of two floats is bigger or even solve really simple logic puzzles for little kids. But I can see how these things mesmerize true believers.
At this point bringing up the ELIZA argument is basically bad faith gaslighting…
Finding bugs in some models doesn’t mean you have a point about intelligence. If somebody could apply a similar argument to dismiss human intelligence, you don’t have a point. And here it goes: the most advanced human intelligence can’t reliably multiply large numbers or recall digits of Pi. Obviously humans are dumber than pocket calculators.
Your counterargument is invalid. The most advanced human intelligence invented (or discovered) concepts like multiplication, pi, etc., and created tools to work around the ways in which these concepts aren't well handled by their biological substrate. When machine intelligences start inventing biological tools to overcome the limits of their silicon existence, you'll have a point.
Designing biological tools is not a commonly accepted bar for AGI.
Isn't the comment you are responding to an example of: "When machine intelligences start inventing biological tools to overcome the limits of their silicon existence, you'll have a point"?
Especially if you remember that the change needed for the first "breakthrough" (GPT4) was RLHF. That is, a model that was specifically trained to mesmerize.
Some of the most advanced AI are tool users and can both write and crucially also execute python, and embed the output in their responses.
As given in a recent discussion: https://chatgpt.com/share/ee013797-a55c-4685-8f2b-87f1b455b4...
(Custom instructions, in case you're surprised by the opening of the response).
While it is true that LLM’s lack agency and have many weaknesses, they form a critical part of what machine learning has lacked until transformers became all of the rage.
The things that LLM’s are bad at are largely solved problems using much simpler technology. There is no reason that LLM’s have to be the only component in an intelligent agent. Biological brains have Specialized structures for specialized tasks like arithmetic. The solution is probably integration of LLMs as a part of a composite system that includes database storage, a code execution environment, and multiple agents to form a goal directed posit - evaluate loop.
I’ve had pretty remarkable success with this architecture running on 12b models and I’m a nobody with no resources.
LLM’s by themselves just come up with the first thing that crosses their”mind”. It shouldn’t be surprising that the very first unfiltered guess about a solution might be suboptimal.
There is a vast amount of knowledge embedded in our cultural matrix, and a lot of that is captured in the common crawl and other datasets.llms are like a search engine for that data , based on meaning rather than semantics.
On the current state of AI - do you believe it has "intelligence" or is the underlying system a "prediction machine"?
What signs do you see that make you believe that the next level (biological intelligence) is on the horizon?
We are but prediction machines https://www.psy.ox.ac.uk/news/the-brain-is-a-prediction-mach...
Wrong. As per the article - a part of our brain is a prediction machine. A human body is more than the sum of its parts.
What does this mean, precisely? How is a human body (or plant, or insect, or reptile/bird/mammal) body ever "more than" it's constituent parts? Wouldn't that violate a conservation law?
https://en.wikipedia.org/wiki/Emergence
Is this a good thing ? Because apparently we’re supposed to be building god. So it sounds like we’re on the wrong track, am I wrong ?
If we’ve just copied our feeble abilities, is that supposed to be exciting?
Is god like intelligent just a prediction machine too ?
Well, if we intend "building god" perhaps we're merely making a copy of a copy of God:
The Sixth Day
Then God said, “Let Us make man in Our image, after Our likeness, to rule over the fish of the sea and the birds of the air, over the livestock, and over all the earth itself and every creature that crawls upon it.” So God created man in His own image; in the image of God He created him; male and female He created them. ”…
Genesis 1:26-27 Berean Standard Bible
We do predictions, but much more important, we are able to create new states. Prediction the classical view assigns probabilities to existing states. What's unique to us and a lot of other biological intelligence is the ability to create new states when needed. This is not implicit in the narrow view of prediction machines
What's the next big step? What will it do? Why do we need or want it? Surely you have the answer.
This means you are sure we are close to automated driving, engineering and hospitality?
We already have "automated driving" in some sense. Some cities have fully autonomous taxi services that have operated for a year or more, iirc.
Nah. We're still not that close. Think of it this way, you turn on an appliance at home and it's what, a 0.0001% chance it will explode in your face? Now automated driving, hospitality etc is all more like a 0.1+% chance something goes wrong still. Huge difference.
I don't really take those taxis as a form of solved automation. It's a nice step though.
Why don't robotaxis count?
Because large, money-rich companies and censored usage is not a proving ground. Amazon "trialled" stores that were automated but ended up being humans. Even without the human factor they weren't proof of successful automation of retail.
We aren't running miles much quicker than 4 mins though. The last record was 3m:43s set by Hicham El Guerrouj in 1999.
The 1-minute mile must be right around the corner, and when that inevitably gets broken, the 1-second mile will follow swiftly.
In fact, humans will be running at relativistic speeds within this century, risking the total destruction of the Earth if they should ever trip over something and fall down.
Scary stuff. And it's not science fiction- it's based on real, observed trends and scaling laws. Seems impossible? Well, they said the four-minute mile was impossible, too.
While this is true, I think you’re not appreciating the metaphor.
Humankind tried to break the 4 minute mile for hundreds of years - since measuring distance and time became accurate enough to be sure of both in the mid-18th century, at least - and failed.
In May 1954, Roger Bannister managed it. By late June it was done again by a different runner. Within 20 years the record was under 3:45, and today there are some runners who have achieved it more than 100 times and nearly 1800 runners who have done it at all.
Impossible for hundred of years, and then somebody did it, and people stopped thinking it was impossible and started doing it themselves. That’s the metaphor: sometimes we think of barriers that are really mental, not real.
I’m not sure that applies here either, but the point is not that progress is continuously exponential, but that once a barrier is conquered, we take on a perspective as if the barrier were never real in the first place. Powered flight went through this. Computing hardware too. It’s not an entirely foolish notion.
For language models specifically, they are trained on data and have historically been improved by increasing the size of the model (by number of parameters) and by the amount and/or quality of training data.
We are basically out of new, non-synthetic text to train models on and it’s extremely hard work to come up with novel architecture that performs well against transformers.
Those are some simple reasons why it will be far more difficult to improve general language models.
There are also papers showing that training models on synthetic data causes “model collapse” and greatly reduces output quality by magnifying errors already present in the model, so it’s not a problem we can easily sidestep.
It’s an easy mistake to see something like chatgpt not exist, then suddenly exist and assume a major breakthrough happened, but behind the scenes there has been like 50 years of R&D that led to it, it’s not like suddenly there was a breakthrough and now the gates are open.
A general intelligence for CS is like the elixir of life for medicine.
this is not even remotely true.
There is an astronomical amount of data siloed by publishers, professional journals etc. that is yet to be tapped.
OpenAI is making inroads by making deals with these content owners for access to all that juicy data.
You seem to think these models haven't already been trained on pirated versions of this content, for some reason.
Are we really now?
The smart people I've spoken to on the subject seem to agree the current technology based on LLM are at the end of the road and that there are no breakthrough in sight.
So what is your take on the next level?
Define breakthrough, there's plenty of room to scale and optimize without any need for a breakthrough (well my definition of breakthrough). Emergent properties so far have been obtained purely from scaling.
There will be no more progress via scaling. All the available training data has already been exploited.
I don't know anything about neuroscience, but is there anything in the brain even remotely like the transformer architecture? It can do a lot of things, but I don't think that it's capable of emulating human intelligence.
I don't know anything about biology, but is there anything in birds even remotely like the airplane? It can do a lot of things, but I don't think that it's capable of emulating bird flight.
Awesome comment. I've written a piece here about the relationship between AI and human consciousness. Would love some feedback if you're able. Thanks! https://peterholmes.medium.com/the-conscious-computer-af5037...
PS. I'm buying your book right now.
There is no guarantee that we will not get stuck with these probabilistic parrots for 50 more years. Definitely useful, definitely not AI.
And by the way I can copy your post character by character, without hallucinating. So I am definitely better than this crop of "AI" in at least one dimension.
Agreed. This is something I didn't think I'd see in my lifetime, let alone be poised to be able to run locally. The alignment is fortuitous and staggering.
People focused on the products are missing out on the dawn of an epoch. It's a failure of perspective and creativity that's thankfully not universal.
What are you talking about? What autonomy? Try the latest Gemini Pro 1.5 and ask it for the list of ten places to visit in Spain. Then ask it for the Google Maps URLs for those places. It will make up URLs that point to nowhere. This os of zero value for personal or business use. I have dozens of examples of such crappy outcomes from all "latest", "most powerful" products. AI is smoke and mirrors. It is being sold as a very expensive solution to a non-existent problem and is not getting any better in the future. Some wish AI had someone like Steve Jobs to properly market it, but even Steve Jobs could not make a crappy product sell. The whole premise of AI goes against what generations of users were told--computers always give correct answers and given the same input parameters return the same output. By extension, we were also taught that GIGO (Garbage-In, Garbage-Out) is what we can blame when we are not happy with the results computers generate. AI peddlers want us to believe in and pay for VIGO (Value-In, Garbage-Out) and I'm sorry but there is not a valid business model where such tools are required.
From a neuroscience perspective , current AI has not helped explain much about real brains. It did however validate the connectionist model of intelligence and memory, to the point that alternate theories are much less believable nowadays. It is interesting to watch the deep learning field evolve, hoping that at some point it will intersect with brain anatomy.
The potential rewards are so great that you might be overestimating the odds this will come about. Even lottery skeptics might buy a lottery ticket if the prize is a billion dollars.
Perhaps it's confirmation bias ?
This looks like a cognitive dissonance and they are addressed by revisiting your assumptions.
No flood-gates have been opened. ChatGPT definitely found uses in a few areas but the number is very far from what many people claimed. A few things are really good and people are using them successfully.
...But that's it. Absolutely nothing even resembling the beginnings of AGI is on the horizon and your assumption that the rate of progress will remain the same -- or even accelerate -- is a very classic mistake of the people who are enthusiasts in their fields.
This is not clear at all. If you know something that nobody else does, please let us know as well.
It's obvious there's potential. It's also obvious it requires at least one other major breakthrough. But no one knows how far away that is.
Why are you throwing in 'consciousness' in a comment regarding mechanical intelligence?
We don’t know what the next big leap will bring and when it will happen. The occurrence of a singular previous big leap cannot serve as any reliable predictor.
The argument that we are going to see massive progress soon is weak in my view. It seems to be:
- we had some big breakthroughs recently
- some AI “godfathers” are “really worried”
Fascinating background - would love to pick your brain on how you see current LLMs/ML comparing to neuroscience. What do you see that's missing still, if anything?
If I had to bet, I would start with:
- Error-correcting specialized architectures for increasing signal-to-noise (as far as I can tell these are what everyone is racing to build this year, and should be doable with just conventional programming systems wrapping LLMs)
- Improved energy efficiency (as yes, human brains are currently much more efficient! But - there are also simple architecture improvements (both software and hardware) that are looking to save 100x. Specialized ASIC ternary chips using 1999's tech should be here quite soon, a lot more efficient in price and energy.)
- No Backwards-propagation. (As yes, the brain does seem to do it all with forward-propagation only. Though this is possible and promising in neural networks like the Forward-Forward algorithm too, they haven't been trained to the same scales as backprop-heavy transformers (and likely have a lower performance in terms of noise/accuracy). Though if I'm not mistaken, the brain does have forward-backward loops, but the signals go through separate neurons for each direction (rather than reusing one) - if so that's close to backprop by itself, but probably imposes a tradeoff as the same signal can't be perfectly reproduced backwards, yet it can perhaps be enhanced to be just the most relevant information by the separate specialized neuron. I'm obviously mostly ignorant of the neuroscience here but halfway-knowledgeable on the ML theory haha
But yes, I completely agree - the flood gates are already open. This is a few architecture quibbles away from an absolute deluge of artificial intelligence that will dwarf (drown?) anything we've known. Good point on decentralized cheap autonomy - the real accomplishment of life. Intelligence, as it appears, is just a fairly generous phenomenon where any autonomous process continually improves its signal-to-noise ratio... many ways to accomplish that one! Looking forward to seeing LLMs powered by ant colonies and slime molds, though I suspect by then there will be far more interesting and terrifying realities unlocked.
I'd like to believe it more than you do. Unfortunately, in spite of these millions of dollars, the progress on LLMs has stalled.
Can you explain what this means? Do you have a degree in neuroscience?