return to table of content

Microsoft strikes deal with Mistral in push beyond OpenAI

philoinvestor
101 replies
23h49m

I don't see how this "AI rush" isn't going to skew markets, skew the financials of big tech and create a bubble in the space.

esics6A
85 replies
23h26m

This is a bubble and one that will burst very hard. AI is the perfect technology for this. It's opaque and most investors (who barely understand tech in general) have no clue what it really does or how it works. This is the closest we've ever had to multiple large respected tech companies selling "snake oil" a cure all. The capabilities of AI they mention as if they're available today are literally many decades and generations away. Automating information workers, creatives and engineers will take AGI that's simply impossible with our technology.

When the AI bubble bursts I wouldn't be surprised if takes down major tech companies with it.

jstummbillig
62 replies
23h16m

Can you be specific about what the bubble is?

I am getting 100x value out of my 30 chatgpt bucks. I am doing things that I could not have done pre-gpt4, being more productive by a factor of, idk, 1.25 maybe.

It's quite simply the largest/simplest productivity improvement in my life, so far. Given it's only going to get better, unless they are underpricing the service by a enormous margin (as in: defrauding shareholders margin) I have a hard time understanding what shape the bubble could possibly have.

PheonixPharts
11 replies
23h8m

You're really generating $3000 per month from ChatGPT? Can you give a hint about what you've built that generates this kind of ROI?

I have only seen people making money in AI by selling AI products/promises to other people who are losing money. The practical uses of these tools still seem to be largely untapped outside of as enhanced search engines. They're great at that, but that does not have a return on value that is in proportion to current investment in this space.

jstummbillig
7 replies
22h48m

Can you give a hint about what you've built that generates this kind of ROI?

Sure. Absolutely nothing amazing: (Mostly) internal software for a medical business I am currently building.

It's just that the actual cost of hiring someone is even quite a bit higher, than what is printed on the paycheck and the risk attached to anyone leaving on a small team is huge (n=0 and n=1 is an insane difference). GPT4 has bridged the gap between being able to do something and not being able to do something at various points over the past year.

EDIT: And to be clear, while I won't claim "rockstar programmer", I have coded for roughly 20 years, which is the larger part of my life.

JumpCrisscross
6 replies
22h22m

Just spoke to a restaurant group owner in Mexico who was able to eliminate their web developer because he can now ask ChatGPT to draft up a basic website.

The kicker? It couldn't do the interactive menu their old website did, so now clicking menu links to a PDF. Which is always, ALWAYS, better.

sgu999
3 replies
22h9m

he can now ask ChatGPT to draft up a basic website.

I'm pretty sure he could have done that with one of the thousands tools like Wix, many years before ChatGPT.

slices
0 replies
21h34m

yes, but which one of thouse thousand, how long would it take to learn how to use it, etc. Still less friction in just asking ChatGPT to do this via the same interface you ask it to do a bunch of other stuff.

phatfish
0 replies
19h32m

Pretty sure he still does that. Unless ChatGPT can now test and deploy a website as well as generate text.

fragmede
0 replies
21h17m

Even just looking at ChatGPT as a better frontend to the Wix help docs, ChatGPT empowered this restaurant owner to do the job themselves, rather than having to have a person do it. Which means that person is out of a job. Good for the restaurant owner, but bad for that person. Which means it's down to personal relationships and how you treat people and all those soft skills that aren't programming.

saewitz
0 replies
21h43m

It's better with accessibility?

dorkwood
0 replies
15h46m

Sorry, but why is pdf better than html? If pdf is better, would you prefer every website just downloaded a pdf to your phone when you visit their url, instead of serving you html? If not, why is it different for a restaurant menu?

doug_durham
1 replies
21h39m

You are only looking at one dimension. What is your hourly rate based on your salary. If ChatGPT saves you 10 hours a month that could easily be over $2000.

spiderice
0 replies
2h49m

But that’s only true if it eventually puts an extra $2,000 in your pocket or an extra 10 hours in your life.

If you estimate that it saves hou 10 hours per month, but your salary stays the same and you don’t work less hours, did it really give you $2,000 in value?

Obviously I don’t know the details of OPs situation. Maybe they aren’t salaries. Maybe the work for themselves. Etc.. I just think people tend to over estimate the value of GPTs unless it is actually leaving them with more money in their pocket.

michaelbuckbee
0 replies
21h36m

This is just a +1 to the ROI discussion, but I'd say that AI tooling roughly doubles my development productivity.

Some of it's in asking ChatGPT: "Give me the 3 possible ways to implement X?" and getting something back I hadn't considered. A lot of it is in sort of "super code completion".

I use Cursor and the UI is very slick. If I'm stuck on something (like a method that's not working) I can highlight it and hit Cmd+L and it will explain the code and then suggest how to fix it.

Hit Cmd+K and it will write out the code for you. Also, gotten a lot of mileage out of writing out a rough version of something in a language I know and then getting the AI to turn that into something else (ex: Ruby to Lua).

wing-_-nuts
10 replies
22h57m

I get that there are limitations with LLMs, but I don't understand people saying it has no value, just because it occasionally hallucinates. Over the past week I've used chatGpt to code not one, but two things that were completely beyond my knowledge (an auto delete js snippet, and a gnome extension that turns my dock red if my vpn turns off). These are just two examples. I've also used it to write a handy regex and write a better bash script.

LLMs are insanely helpful if you use them with their limitations in mind.

aleph_minus_one
7 replies
22h43m

LLMs are insanely helpful if you use them with their limitations in mind.

This depends on your use case. I can honestly tell that all the chat bot AIs don't "get" my kind of thinking about mathematics and programming.

Since some friend who is graduate student in computer science did not believe in my judgement, I verbally presented him some test prompts for programming task where I wanted the AI to help me (these are not the most representative ones for my kind of thinking, but are prompts for which it is rather easy to decide whether the AI is helpful or not).

He had to agree from the description alone that the AIs will have difficulties with these task, despite the fact that these are common, and very well-defined programming problems. He opined that these tasks are simply too complex for the existing AIs, and suggested that if I split these tasks into much smaller subtasks, the AI might be helpful. Let me put it this way: I personally doubt that if I stated the subtasks in a way in which I would organize the respective programs, the AI would be of help. :-)

What was just important for me was to able to convince the my counterpart that whether AIs are helpful or not for programming depends a lot on your kind of thinking about programming and your programming style. :-)

mewpmewp2
2 replies
20h52m

Any chance you could share the prompt?

I on the otherhand feel like I am completely in sync with Copilot and ChatGPT. It is as if it always knows what I am thinking.

aleph_minus_one
1 replies
15h29m

Don't take the following prompts literally, but think into the directions of:

"Create a simple DNS client using C++ running on Windows using IO Completion ports."

"Create a simple DNS client using C++ running on GNU/Linux using epoll."

"Write assembler code running on x86-64 running in ring 0 that sets up a minimal working page table in long mode."

"Write a simple implementation of the PS/2 protocol in C running on the Arduino Uno to handle a mouse|keyboard connected to it."

"Write Python code that solves the equivalence problem of word equivalence in the braid group B_n"

"Write C++|Java|C# code that solves the weighted maximunm matching matching problem in the case of a non-bipartite graph"

...

I experimented with such types of prompts in the past and the results were very disappointing.

All of these are tasks that I am interested in (in my free time), but would take some literature research to get a correct implementation, so some AI could theoretically be of help if it was capable of doing these tasks. But since for each of these tasks, I don't know all the required details from memory, the code that the AI generates has to be "quite correct", otherwise I have to investigate the literature; if I have to do that anyway, the benefit that the AI brings strongly decreases.

mewpmewp2
0 replies
10h57m

I tried the first one, but currently I don't have time to verify what it generated.

But I have done many Arduino/Raspberry PI things lately for the first time in my life and I feel like ChatGPT/Copilot has given me a huge boost even if it doesn't always give 100 percent code out of the box, it will give me a strong starting point where I can keep tweaking myself.

wing-_-nuts
1 replies
22h39m

I would say that the ability to break a problem down into manageable chunks is the mark of a sr dev. I think of chatGpt as a jr that's read a lot but understands only a little. To crib Kurtzwell you gotta 'run with the machine'

aleph_minus_one
0 replies
15h45m

I would say that the ability to break a problem down into manageable chunks is the mark of a sr dev.

I believe that I am perfectly capable of doing this. But if I have to "babysit" the LLM, its helpfulness decreases.

nuancebydefault
0 replies
20h40m

This is a rather long post, I'm genuinely curious why you did not describe the problem that you want to solve. Is it too complex for even humans to understand?

fragmede
0 replies
21h23m

What did happen when you split the question into subtasks? What were those questions?

chasd00
1 replies
20h32m

LLMs are insanely helpful if you use them with their limitations in mind.

the fact that LLM responses can't be add supported (yet) make them much more valuable than internet search IMO. You have to pay for chatgpt because there's no ads. No ads no constant manipulation of content and your search to get more ads in front of you.

Having to pay for using genai is it's best selling point ironically.

oblio
0 replies
14h59m

This is only temporary. Ads will be put back in as soon as possible.

nozzlegear
8 replies
22h50m

Last night I asked ChatGPT 4 to help me write a quick bash script to find and replace a set of 20 strings across some liquid files with a set of 20 other strings. The strings were hardcoded, it knew exactly what they were in no unclear terms. I just wanted it to whip up a script that would use ripgrep and sed to find and replace.

First, it gave me a bash script that looked pretty much exactly like what I wanted at first glance. I looked if over, verified it even used sed correctly for macOS like I told it, and then tried to run it. No dice:

    replace.sh: line 5: designer_option_calendar.start_month.label: syntax error: invalid arithmetic operator (error token is ".start_month.label")
Not wanting to fix the 20 lines myself, I fed the error back to ChatGPT. It spun me some bullshit about the problem being the “declaration of [my] associative array, likely because bash tries to parse elements within the array that aren’t properly quoted or when it misinterprets special characters.”

It then spat out a “fixed” version of the script that was exactly the same, it just changed the name of the variable. Of course, that didn’t work so I switched tactics and asked it to write a python script to do what I wanted. The python script was more successful, but the first time it left off half of the strings I wanted it to replace, so I had to ask it to do it again and this time “please make sure you include all of the strings that we originally discussed.”

Another short AI example, this time featuring Mistral’s open source model on Ollama. I’d been interested in a script that uses AI to interpret natural language and turn it into timespans. Asking Mistral “if it’s currently 20:35, how much time remains until 08:00 tomorrow morning” had the model return its typical slew of nonsense and the answer of “13.xx hours”. This is obviously incorrect, though funnily enough when I plugged its answer into ChatGPT and asked it how it thought Mistral may have come to that answer, it understood that Mistral did not understand midnight on a 24 hour clock.

These are just some of my recent issues with AI in the past week. I don’t trust it for programming tasks especially — it gets F# (my main language) consistently wrong.

Don’t mistake me though, I do find it genuinely useful for plenty of tasks, but I don’t think the parent commenter is wrong calling it snake oil either. Big tech sells it as a miracle cure to everything, the magic robot that can solve all problems if you can just tell it what the problem is. In my experience, it has big pitfalls.

codexon
4 replies
22h20m

I have the same experience. Every time I try to have it code something that isn't completely trivial or all over the internet like quicksort, it always has bugs and invents calls to functions that don't exist. And yes, I'm using GPT-4, the best model available.

And I'm not even asking about an exotic language like F#, I'm asking it questions about C++ or Python.

People are out there claiming that GPT is doing all their coding for them. I just don't see how, unless they simply did not know how to program at all.

I feel like I'm either crazy, or all these people are lying.

nkozyra
1 replies
21h37m

I feel like I'm either crazy, or all these people are lying.

With some careful prompting I've been able to get some decent code that is 95% usable out of the box. If that saves me time and changes my role there into code review versus dev + code review, that's a win.

If you just ask GPT4 to write a program and don't give it fairly specific guardrails I agree it spits out nearly junk.

oblio
0 replies
14h45m

If you just ask GPT4 to write a program and don't give it fairly specific guardrails I agree it spits out nearly junk.

The thing is, if you do start drilling down and fixing all the issues, etc, is it a long term net time saver? I can't imagine we have research clarifying this question.

throwawayanothe
0 replies
21h40m

Same experience here

taco-hands
0 replies
20h13m

People are out there claiming that GPT is doing all their coding for them. I just don't see how, unless they simply did not know how to program at all.

I doubt it and certainly not for anything beyond basic. I've seen (and tried GPT's for code input a lot) and often they come back with errors or weird implementations.

I made one request yesterday for a linear regression function (yes, because I was being lazy). So was chatGPT... It spat out a trashy broken function that wasn't even remotely close to working - more along the lines of pseudo code.

I complained saying "WTH, that doesn't even work" and it said "my apologies" and spits out a perfectly working accurate function! Go figure.

Others have turned to testing tips or threats, which is an interesting avenue: https://minimaxir.com/2024/02/chatgpt-tips-analysis/

jstummbillig
2 replies
22h16m

I hear you. It's all pretty bad. I have spent half-days getting accustomed to and dealing with gpt garbage — but then I have done that plenty of times in my life, with my own garbage and that of co-workers.

On the margins it's getting stuff good enough, often enough, quick enough. But it very much transformed my coding experience from slow deliberation to a rocket ride: Things will explode and often. Not loving that part, but there's a reason we still have rockets.

oblio
1 replies
14h44m

Most of us don't ride nor want to ride rockets. We do use them to blow each other up, statistically.

jstummbillig
0 replies
4h58m

Most of us don't ride nor want to ride rockets.

The amount noise generated by pretty much anything new and shiny on this website would disagree with that.

We do use them to blow each other up, statistically.

Very true — and yet :^)

rchaud
6 replies
22h55m

The bubble is that every $1 in capital going to OpenAI/Nvidia is a $1 that cannot be invested anywhere else: Healthcare, Automotive, Education, etc. Of course OAI and Nvidia will invest those funds, but in areas beneficial purely to them. Meta has lost $20bn trying to make Horizon Worlds a success, and appears to have abandoned it.

Even government-led industrialization efforts in socialist economies led to actual products, like the production of the Yagan automobile in Chile in the 1970s[0].

We've already had a decade plus of sovereign wealth funds sinking tens of billions into Uber and autonomous driving. We still don't have those types of cars on the road and it's questionable whether self driving will even generate the economic growth multiplier that its investment levels should merit.

[0] https://journal.hkw.de/en/erinnerungen-an-den-yagan-allendes...

polski-g
2 replies
21h43m

When Meta "lost" 20bn, they actually spent it on salary. The employees then go out and buy things like the Tesla Automobile, an actual product.

rchaud
0 replies
21h24m

If the same $20bn was spent on fixing a bridge, people would spend those wages to boost economic activity AND have a fixed bridge that will improve output even more. Horizon Worlds isn't a productive use of capital in that regard.

It'd be one thing if they open-sourced their VR tech, some of that could lead to productive tech down the line, but as a private company, they're not obliged to do any of that.

acdha
0 replies
21h40m

This sounds like the broken window fallacy. You could use the same logic to suggest that Meta dump piles of cash on the sidewalk in front of their office - it’d circulate but it wouldn’t help them.

oblio
0 replies
14h48m

I did a quick check and it seems like the entire Uber and clone industry is net negative. Uber, Lyft, Didi, Grab seem to have lost more money than was invested and once they stabilize they look like mediocre businesses at a global scale (Uber's been banned from many jurisdictions for predatory practices and in many other jurisdictions it seems to trend towards being as expensive as taxis or more once profitability becomes a target).

jstummbillig
0 replies
22h28m

Fair enough. I personally would have a hard time spotting an outsized lost opportunity value with confidence, if it existed.

It feels, though, that this argument could (maybe a be little too easily) be applied to any new industry sector, in horse-vs-car fashion.

DANmode
0 replies
22h42m

As well as the artificially increased valuations for every company with the .ai TLD for their landing page.

spiderice
5 replies
23h5m

I’m really curious what is making you so much more productive. My experience with AI has largely been the opposite. Also curious how you’re using AI to make $3,000 per month more than you would without it.

choilive
3 replies
21h45m

It's not that he is making $3k a month, he gaining the equivalent productive value of of $3k.

oblio
2 replies
14h55m

How is he measuring that, though?

choilive
1 replies
12h10m

Average time saved per task * Number of tasks * Cost of their time per hour.

At $100/hr (not unreasonable for Sr. SWE) - he just needs to save 30 hrs/mo.

Not to mention the opportunity cost of using that time on more impactful activities on their startup. AI can be a force multiplier for sure.

spiderice
0 replies
3h0m

Then he is either making an extra $3,000/month or working 30 hours less per month. If the former isn’t true, I am extremely doubtful that the latter is.

Seems more likely that he is over estimating the value that LLMs are bringing him. Or he is an extreme outlier, which is why I was asking for further details

aquova
0 replies
20h49m

I feel the same way. I think LLMs are neat, and I find them interesting from a technical standpoint, but I have yet to have them do anything for me that's more than just a novelty. Even things like Copilot, which I'll admit has impressed me a bit, doesn't feel like it would radically change my life, even if it was completely foolproof.

philoinvestor
3 replies
22h28m

You may be getting 100x value out of your 30 bucks.

Is that the only variable that one needs to consider to gauge if this is bubble territory?

jstummbillig
2 replies
22h26m

Other variables is basically what I was asking for, in the first sentence of my comment.

philoinvestor
1 replies
22h3m

How many GPUs are being delivered today and for how long will they be used / what's their life?

Who is funding the purchase of those GPUs?

If VC money then what happens if the startups don't make money?

Are users using AI-apps because they are free and dump them soon?

Isn't their competition in semiconductors? Won't we have chips-as-a-commodity soon? LLMs-as-a-commodity?

Is Big Tech spending all this money to create VALUE or just to survive the next phase of the technological revolution? (e.g. the AI rush)

If prices are high, and sales are high, and competition is still low -- then how much is nvidia actually worth? And if we don't now why is it selling for so many times earnings?

https://www.philoinvestor.com/p/downside-at-nvidia-and-the-n...

jstummbillig
0 replies
3h45m

These are all potentially interesting questions, but do any of them offer a concise answers to the bubble-question or are we just speculating?

dottjt
3 replies
22h4m

It's that "it's only going to get better" part that is driving the bubble, I think.

The market has this idea that over the next year, we're somehow going to have AI that's literally perfect. Yet that's not how technology works, it takes decades to get there.

It'd be like if the first LCD TV was invented, and all of a sudden everyone is expecting 8k OLED by the next year. It just doesn't work like that.

thatoneguy
1 replies
21h13m

Yeah, but this is different because it's largely just money -> more GPUs -> more people looking at the problem -> better results. You can't stumble upon an 8K TV overnight but you can luck upon the right gold mining claim and you can luck upon some new training algorithm that changes the game immediately.

oblio
0 replies
14h56m

This assumes answers which go from 8/10 to 9/10 increase the cost linearly. The scale could be exponential or worse.

jstummbillig
0 replies
11h12m

Fair – but again: If it just stayed frozen in the state it is now (and that assumption is about as unreasonable, as it being "perfect" in a year), it's already going to be tremendously useful for increasingly many people (when cost will go down, and they will, accessibility will go up) — at least until something better comes around.

For those who extract value right now, the simple alternative (just not using it) is never going to be the better choice again. It's transformative.

philoinvestor
2 replies
22h31m

Can I paste links in replies here? I wrote an article about the AI bubble a few weeks ago.

nozzlegear
1 replies
19h0m

Yes you can.

Judgmentality
2 replies
23h11m

Do any AI companies actually turn a profit? I feel like the only real winner is Nvidia because they are selling shovels to the gold diggers, while all the gold diggers are trying to outspend each other without a business model that has any consideration for unit economics.

jstummbillig
0 replies
23h5m

I love a prudent take on company money – but given how investing works and how young this entire thing is and the (to me) absolutely real value, I find it hard to be very worried about that part right now.

I can literally run a ballpark model on my MB Pro, right now, at marginal additional electrical cost. I will be the first to say that all of this (including GPT4) is still fairly garbage, but I don't know when there was the last time in the history of tech, where less fantasy to get from here to what will be good was required.

chucke1992
0 replies
23h0m

The thing is that the bigger business giants like MSFT or Amazon are probably profit quite nicely from AI. Smaller companies, not aligned with any big giants - probably not.

outside1234
0 replies
22h41m

There is very clear value in LLMs

The bubble is that it is not clear there is a $XXXB business in building or hosting them.

OpenAI is losing money hand over fist, open source models are becoming available that are on-par and so commoditize the market, etc.

novaRom
0 replies
22h40m

largest/simplest productivity improvement in my life, so far

many productivity improvements in the last years: Internet Search, Internet Forums, Wikipedia, etc. LLMs and other AI models is continuation of the improvement of information processing.

jononor
0 replies
18h59m

ChatGPT plus is approx 300 USD yearly. You are earning 30k (100x) USD more yearly due to ChatGPT? Or earning 100k with 30% use or your time?

SunlitCat
5 replies
22h3m

It's like the previous craze with blockchain. Everyone and their dog was about to use blockchain to do the most awesome thing ever. And even things that didn't really lend themselves to blockchain were suddenly presented as prime examples of how to use blockchain.

I'm not saying both technologies don't have their uses, but the hype around them is crazy and not healthy.

hackerlight
2 replies
19h2m

AI has massively increased my productivity. There are real users like me who will continue to pay up. This is not the same.

oblio
0 replies
14h42m

Yeah, but how many people like you are out there? We'll find out in about 5 years.

SunlitCat
0 replies
14h5m

Like with blockchain, there are real uses, of course i totally agree!

But I was more thinking about the craze surrounding the whole thing, like with blockchain, you can see everyone trying to sell you AI for kinda everything.

philoinvestor
1 replies
20h6m

Yep, there are similarities between the blockchain rush and the AI rush.

But in my last piece on AI, I said that AI is 50X Blockchain!

SunlitCat
0 replies
12h32m

That answer inspired me to ask ChatGPT on how blockchain could enhance AI. And, to say the least, I wasn't disappointed!

It printed out a pretty long answer with several good points on how it could help to enhance AI.

There we have it! Blockchain is about to solve all problems we have today with AI! :D

d--b
3 replies
21h36m

Well that's your view, but we keep seeing incredible progress year on year.

Noone thought beating the game of Go was feasible.

Noone thought self-driving cars would actually work.

Noone predicted ChatGPT. See where we're at now with multi modal models.

And Sora.

The truth is that you can't predict anything anymore about this tech cause all expectations keep being blown away.

AGI may be right there, and that's what's driving the money.

oblio
0 replies
14h39m

Noone thought beating the game of Go was feasible.

Oh, yes, we did, once we beat chess. It was just a matter of time.

Noone thought self-driving cars would actually work.

And... they don't? Call me when I can buy a regular car where I can sleep while traveling 8 hours, driven by the car itself. We're probably "flying cars" away from that.

SirMaster
0 replies
16h48m

This doesn’t make any sense. If nobody thought the things were possible then why did they bother trying to do them?

And even spend as much time and resources as they did trying to do them?

ArchOversight
0 replies
21h28m

Noone thought self-driving cars would actually work.

Many people thought self driving WOULD work and that we'd be further along than we are now. We have vastly overestimated how far we'd be, and vastly underestimated how much time and effort it would actually take.

Self driving cars as they exist today are still mere toys compared to where the industry thought they were going to be. Look at Cruise, Waymo, Zoox, Uber's ex-self driving car division and others.

We are not anywhere near the self-driving autonomous cars we had hoped for.

codexon
3 replies
22h0m

This technology is already fully automating copywriters and almost replacing concept artists right now. However it is true that right now the valuations are for something far more than that, and so far it doesn't seem like LLMs will be able to do much more.

I was talking to someone that just retired from a programming position at a FANG and he seems to think that AGI (artificial general intelligence) is only a few years away just based off what he sees with ChatGPT and he's dumping all his money into AI stocks. The level of hype and over-extrapolation is so absurd, and the fact that it can affect someone with a technical background...

It really does seem like a bubble to me.

ForgotIdAgain
2 replies
21h53m

Isn't the role of a concept artist mainly to do worldbuilding and drawing second? AI does not seem to have a good world model, they make pretty pictures but they lack thought behind them.

codexon
0 replies
21h38m

I really don't know the details.

You can see a concept artist here discovering he stopped getting work after a company splurted out that they switched to AI.

https://twitter.com/_Dofresh_/status/1709519000844083290

My guess is that a lot of people can have ideas, so you don't need an artist to bring them to life anymore.

aquova
0 replies
20h44m

Agreed, and I think there are a number of... over-enthusiastic executives with dollar signs in their eyes who are in for a rude awakening about this. It might sound great to replace your artistic staff with an LLM subscription, until you realize that you laid off all the creative vision with them. That isn't to say I think it'll go away though, I wouldn't be surprised if art students in the future are taught how to wrangle LLMs to supplement their own designs.

philoinvestor
1 replies
22h30m

I agree that it's opaque and that's why I said that AI can be the ULTIMATE bubble.

mnky9800n
0 replies
22h28m

It's a shame the phrase "black box" exists in machine learning when it could have been "black bubble".

partiallypro
0 replies
22h15m

I think there is a bubble, but not in the traditional sense of this all being nonsense. Obviously, AI is going to be a world changer, but I think in a few years many companies are going find, at least for them, that their projects using AI are wildly unprofitable. Good AI is expensive. This could take years to play out, but when Nvidia is trading higher than Microsoft and Apple in total market cap, we'll be well into it. I don't rule that out.

huytersd
0 replies
20h16m

Eh I don’t know. I find ChatGPT very useful atleast 10 times a day.

dorkwood
0 replies
13h8m

This opinion is so crazy to me. How can it be a bubble if it's already providing me with 10x productivity gains? Things I've used GPT-4 to automate: replaying to all emails, writing entire books, doing all the legal paperwork for my new business, completing daily code challenges, generating shopping lists, writing birthday messages, and replying to my family group chat. And that's when the tech is still relatively new! Once the technology advances and I'm able to automate basically all of my daily tasks, who knows where I'll be.

curtisblaine
0 replies
21h52m

When the AI bubble bursts I wouldn't be surprised if takes down major tech companies with it.

Hopefully, knock wood. Maybe it will even slow down the genrral enshittification created by those major tech companies.

Spivak
0 replies
22h26m

This is the closest we've ever had to multiple large respected tech companies selling "snake oil" a cure all.

The problem with this take is you can deliver real results. At current $dayjob we do the very dumbest thing which is text -> labels -> feedback -> fine_tune -> text... and surface them as part of our search offering and it's rocketed to the most useful customer feature in less than 6 months of rolling it out. Customers define labels that are meaningful to them, we have a general-purpose AI classify text according to those labels. Our users gleefully (which is shocking given our industry) label text for us (which we just feed into fine_tuning) because of just how fast they can see the results.

Like it's as grug brain as it gets and we bumbled into a feature that's apparently more valuable to our users than the rest of the product combined. Folks want us to sell it as a separate module and we're just hoping they don't realize it's 3 LLMs in a trenchcoat.

ABS
0 replies
22h23m

P/E ratios (so far) say otherwise: all the big tech public companies have reasonable P/E ratios and are investing heavily their profits in contrast to, say, the dotcom bubble when over 40% of the companies had P/E ratios unsustainably over the moon.

Are there a gazillion companies riding the "AI everywhere" wave to raise money? yes, yes there are. Will most of them fail? sure.

But the big players are fine at the moment so there is nothing that can burst very hard (yet) and the difference is in the denominators which are, so far, going up.

Of the top ones only NVIDIA and Amazon have P/E ratios a bit too high and among the top 10 only AMD's is way too high.

gitfan86
8 replies
23h43m

There is a bubble in the valuations of LLMs similar to the valuation of NFTs. Anyone can create a modified version of Llama2.

thethimble
5 replies
23h36m

The last All in Podcast had an interesting take on this:

Most of the R&D and Capex going into LLMs/GenAI is speculative. The investments haven’t translated into real revenue yet. The expectation is that there will be a large pot of revenue at the end of the road, but we haven’t seen the killer apps to substantiate this. This makes for a perfect bubble if the promise doesn’t pan out.

Relatedly Nvidia’s revenue - as impressive as the recent growth has been - is fully exposed to this risk.

Of course it’s possible (likely?) that there will be major wins from this tech, but the fact that there isn’t definitive proof (in the form of revenue) yet represents real risk.

pantulis
1 replies
23h29m

The expectation is that there will be a large pot of revenue at the end of the road, but we haven’t seen the killer apps to substantiate this

Imho, unlike crypto and NFT, AI is not a solution in search of a problem. While there is a lot of hand waving, it is not very adventurous to predict that there will be significant productivity gains by adding AI on top of current business processes. Thus the killer apps will be... the same apps we are using today with a touch of AI magic dust.

mnky9800n
0 replies
22h26m

Also, outside of making chatbots and waifus there's a lot of hard tech problems waiting for ai type solutions.

chasd00
0 replies
20h19m

that's the root of my skepticism. I've yet to see huge transformational projects involving AI from the grapevine of my 750k+ employee consulting firm, all I hear about are experiments, webinars, and POCs. What I see in the news regarding the few projects live is not good, AirCanada and some others where the AI went wrong and now the company has to backtrack.

There's no doubt the tech community is all excited because genai, indeed, helps write code but I've yet to hear a large company like Coca Cola announce large AI transformation projects the way they announced large cloud transformation projects a few years ago.

I get more and more on the AI bandwagon as time goes on but I still have a pretty healthy skepticism on how deep and wide the tech will penetrate day to day business at enterprises and that's where the ROI is.

andai
0 replies
23h29m

Well, the major win will be when AI makes human labor uneconomical. Not sure what that's going to do to the stock prices though...

JumpCrisscross
0 replies
22h19m

All in Podcast

Calacanis and Palihapitiya are the Jim Kramers of tech.

ctrw
0 replies
23h38m

Yes and each of those costs between half a million in pure compute to a few hundred million for fine tuning/training.

codexon
0 replies
21h58m

The open source models are VERY far behind gpt4.

PheonixPharts
4 replies
23h11m

I've stopped believing in rational markets a long time ago. For the people behind this funding madness:

to skew markets, skew the financials of big tech and create a bubble in the space.

Are the intended consequences of this. The people behind the money in AI, just like the people behind the money in crypto, don't care if there's a reality to all of this, they just care if they can make a lot of money while the music is playing.

I really thought 2022 was going to be the beginning of tech returning to reality, but naively didn't understand that this would entail a lot of people with a lot of money losing money, and that's not going to happen.

As with all bubbles, the interesting thing isn't pointing out there's a bubble, we've been living in many bubbles for decades now. The interesting thing is pointing out what will make it pop. So long as globally money keeps pouring into US markets we'll see this continue.

On the plus side, at least LLMs are a lot of fun to work and play with!

throwup238
0 replies
23h2m

> I've stopped believing in rational markets a long time ago.

It makes a lot more sense if you think of markets as just a bunch of coked out hairless monkeys. Then "rational" becomes a lot more meaningful.

rchaud
0 replies
23h4m

"Efficient markets" deliver normal returns. The people with the most capital to throw at things are trying to create monopolies that can generate supernormal returns.

This is already a highly unstable arrangement, and it's made more dangerous by introducing impurities like artificially suppressed interest rates and SPAC IPOs.

philoinvestor
0 replies
22h26m

With the way NVIDIA is moving and how it's seeding AI startups, you could be right about them actively trying to do this. But there is a reality beyond our ability to predict it..

DANmode
0 replies
22h39m

I've stopped believing in rational markets a long time ago.

Agreed.

Sentiment is everything.

tmpz22
0 replies
23h29m

Theres a lot of not-great software that can have bites taken out of it by LLM advancements and most of the incumbents are going to pay into it because they just can't compete in the AI field.

Can LLMs become reliable enough at transforming data that it can replace or augment the current slew of ETL tools? Can it produce visualizations better then current BI tools? Can copilot compete with a junior developer? Im not sure, but at this point Im willing to say 50/50 which is worth the bet.

bhouston
82 replies
1d3h

This is just a smart move by Satya Nadella after the non-standard drama that occurred with OpenAI a few months back where it nearly imploded and then didn't.

You want both a backup for OpenAI as well as negotiating leverage if OpenAI gets too powerful and this achieves both.

Xenoamorphous
41 replies
1d3h

I guess it has a cost, though? I presume OpenAI didn’t like this move. If that’s the case, what might be the consequences?

bhouston
32 replies
1d3h

I guess it has a cost, though? I presume OpenAI didn’t like this move. If that’s the case, what might be the consequences?

Until OpenAI releases GPT 5 and it blows everyone away, OpenAI's leverage is constantly decreasing as the gap between their best model and everyone else's best model decreases.

There doesn't seem to be moats right now in this industry except for pure model performance.

Maybe someone should as ChatGPT what OpenAI should do to maintain long-term leadership in this industry?

robrenaud
13 replies
1d2h

If I had to pick one player who wanted to win the AI race and was willing to be ruthless to do it, I'd pick Nvidia. Computation is the excludable bottleneck, and Nvidia is the essentially the singular company who makes AI computers.

Hire Ilya, get him to hire as many of the best folks he can.

Stop selling GPUs. Hoard them. Introduce some subtle bug into the drivers that dramatically increases their rate of burn out.

Figure out some reasonable way to give attribution to original content creators, approximately solve the content ID problem of the AI age. Cut the content creators into the rev share in proportion to their data importance to the model. Make the content creators incredibly pissed off that their work is being stolen by big AI companies unfairly and encourage to them to sue the other big AI firms. Their content share multiplier increases if they get injunctions against LLM firms.

Convince politicians that the AI firms have performed an intellectual heist of epic proportions, and that they must not be allowed to even generate synthetic training data from poisoned models. With the content creators united behind you, convince congress that poisoned models must be destroyed, that even using synthetic training data from poisoned models must be illegal. Make them start over from a clean room with no copyrighted data.

ben_w
4 replies
1d

Make them start over from a clean room with no copyrighted data.

And when such models become popular[0], all the artists now have no job and no way to get compensation for being unable to work through no fault of their own.

I don't think that's really a winning condition. It might make you feel better about the world, but the end result is still all the artists being out of work.

[0] some models are already trained that way, although I assume you're using the word "copyrighted" in the conventional sense of "neither public domain nor an open license", as e.g. all my MIT licensed stuff is still copyrighted but it's fine to use.

rusticpenn
1 replies
9h2m

It might make you feel better about the world, but the end result is still all the artists being out of work.

Is it not the same with every leap in technology? There were professions like street lamp lighters, alarm services etc that have become redundant now?

ben_w
0 replies
1h15m

Sure, though I suspect "art" is the human version of a peacock tail — the difficulty is the point, it how we signal our worth to others, cheapening it breaks that signal — which would suggest that making all forms of art easy messes with (many of) us at a deep, essentially automatic, level.

More specifically, I was responding to the idea that "compensating creators whose works are used to train the models" would actually solve anything; to use your examples, it would be as if the literal luddites were suggesting passing laws saying that "all textile machines that work like humans need to compensate the humans they displace, and also you need to make your new machines from scratch without talking to any textile workers to make sure you don't cheat", and my response would be analogous to saying "there's already machines which don't work like humans, so you're going to be out of work and have no compensation".

The Luddite movement preceded The Communist Manifesto by about 30 years. Everything's sped up since then, so I'd be surprised if we have to wait 30 years for a political shift which is to AI what Communism was to industrialisation. I'm just hoping we don't get someone analogous to Stalin or Pol Pot this time.

robrenaud
1 replies
23h34m

In my hypothetical future, at least the people who create the content used to train the models can get "training royalties", which they aren't getting now.

There is still also money to be made in producing physical art or performances, even when AI can produce amazing digital works.

ben_w
0 replies
10h37m

"Make them start over from a clean room with no copyrighted data." makes "the people who create the content used to train the models" the empty set.

There is still also money to be made in producing physical art or performances, even when AI can produce amazing digital works.

Perhaps, but it may be akin to the way there is still money to be made from horse drawn carriages in city centres, even when cars displaced them over a century ago — a rare treat for special occasions, to demonstrate wealth.

johngossman
2 replies
1d1h

Stop selling GPUs. I'm not sure Wall St. would reward that plan
smegger001
0 replies
16h6m

don't stop just raise the margin slightly and limit the number available of the higher end chip using the proceeds to self fund building their own datacneter

do runs of cards for themselves with higher core counts and clock speed that they dont release to others.

CamperBob2
0 replies
1d1h

Yeah, he seems to have left out the critical middle step ("Gather investors to take $2T company private.")

codexon
1 replies
22h35m

If google benchmarks are to be believed, gemini 1.5 will be better than gpt and they use their own chips (Google TPU), no nvidia involved. There is also Groq. I don't see Nvidia keeping their lead and profit margins forever.

vGPU
0 replies
13h45m

Not to mention intel rapidly catching up in the $/perf calculation.

bhouston
0 replies
1d

If I had to pick one player who wanted to win the AI race and was willing to be ruthless to do it, I'd pick Nvidia. Computation is the excludable bottleneck, and Nvidia is the essentially the singular company who makes AI computers.

I've thought the same thing. NVIDIA getting into AI seriously is a vertical integration play and they often do that -- like NVIDIA trying to buy ARM.

SunlitCat
0 replies
22h7m

Stop selling GPUs. Hoard them. Introduce some subtle bug into the drivers that dramatically increases their rate of burn out.

Well, they didn't stop selling GPU when cryptomining was going strong. Instead they continued to sell them (with a hefty markup, tho).

It's like selling shovels and picks during a gold rush.

Ringz
0 replies
1d1h

Sounds like someone has finally asked ChatGPT for a feasible plan to consolidate the AI landscape.

asynchronous
12 replies
1d3h

They also have brand recognition, for what it’s worth. Every non-engineer in the world practically thinks AI == ChatGPT.

ethanbond
8 replies
1d2h

Meh, I don't think it's worth much. In a few years that'll be like claiming that so-and-so had name brand recognition for transistors. Most people don't need to care who manufactures their transistors.

ben_w
3 replies
1d

Brands can change quickly, but they do matter in the short term. I've witnessed customer support teams use Firefox to say they only supported Internet Explorer and government ministers who thought it was "good" that IE was the "only" web browser, and weirdly a phone company whose customer support person thought their SIM cards worked better on Android than iPhone and that their web chat wouldn't work with a Mac even though they were talking to me on a Mac at the time.

And when I was a kid, it seemed like all the teachers thought it would be a waste of time to learn MacOS because "Apple would be bankrupt soon". (Given how much all the app UIs changed, right decision for the wrong reason).

ethanbond
2 replies
23h47m

All of these examples are end-products. "AI" itself will not be. The winner in AI will be whoever permeates other products/brands most successfully, and end-user brand familiarity doesn't matter much for that. Familiarity among engineering and product leaders is what matters.

mitjam
0 replies
22h33m

Maybe, but maybe AI will become front and center of consumer and productivity IT products and their premier brand ambassadors will be anthropomorphized AI agents. Hello Clippy, this time for real.

ben_w
0 replies
22h56m

Indeed, I'm not disagreeing on that, merely opining that "ChatGPT" as a name could well be relevant for a bit.

donny2018
1 replies
1d2h

It’s not necessarily a bad thing. Most people don’t know that TSMC exists, or what Microsoft does beyond Windows and Xbox (which are a small fraction of its business).

ethanbond
0 replies
23h48m

Correct, I didn't say it's a bad thing. I said it's not clear that it's a good thing (i.e. an asset)

bhouston
0 replies
1d2h

Unless your market is direct to end user, end user brand name recognition doesn't matter. In the case of AI, at least so far, the primarily income won't be from end-users directly, but rather via enterprise integrations into existing tools that already have end user market share (e.g. Microsoft Office, Microsoft Windows, VS Code, Notion, etc.)

CamperBob2
0 replies
1d1h

Most people don't need to care who manufactures their transistors.

They might, in an upside-down world where the Shockley Semiconductor board tried to fire Shockley, and where the Traitorous Eight not only didn't bail out but took his side.

vinay427
0 replies
1d2h

Yes, however it's increasingly likely that the GPT in ChatGPT will not be limited to OpenAI (in the US), so I'm not sure how much ChatGPT will be worth with countless other platforms containing GPT in their names.

https://techcrunch.com/2024/02/15/no-gpt-trademark-for-opena...

bhouston
0 replies
1d2h

Sure that helps with the consumer market, but most people will use AI integrated into other products and not directly.

Those integrated AI solutions will usually be done via enterprise deals where brand name is not quite as important. It will be done by people who care about cost, reliability and ease of use.

Think of nginx's dominance in web servers even though it has no name recognition among the general population. Or Stripe's payment system.

YetAnotherNick
0 replies
1d2h

The thing is that there is almost no lock in in the models. So brand recognition doesn't help much as people look into the benchmarks and price sometime in the future, if not when just starting out.

behnamoh
2 replies
1d2h

There doesn't seem to be moats right now in this industry except for pure model performance.

Hard disagree. OpenAI's function calling is something no other commercial model provides, not even Gemini and Mistral Large.

fzzzy
0 replies
1d2h

The mixtral large marketing at the top of hn right now claims json mode and native function calling. I haven't tried it, but that's what they say.

jorvi
0 replies
1d2h

There doesn't seem to be moats right now in this industry except for pure model performance.

Compute?

At least in the short term, it seems like the biggest wallets are going to win by default.

ben_w
0 replies
1d

Eh, all this talk of "moats" etc. feels weird when just a few years ago it seemed like everyone was complaining they'd rearranged their corporate structure to include a fully-owned profit-making subsidiary to attract investments, and all the loud voices seemed to think a cap of x100 return on investment was so large it was unlikely to be reached.

And then OpenAI tripped and fell over a magic money printing factory, and the complaints are now in the set ["it's just a stochastic parrot", "it's so good it's a professional threat to $category", "they've lobotomised it", "they don't have a moat", "they're too expensive"].

As the saying goes, "Prediction is very difficult, especially if it’s about the future!"

iambateman
2 replies
1d2h

I suppose they didn't but Microsoft has a $3T market cap and OpenAI is theoretically valued at $80B.

OpenAI has 700 people, Microsoft has 220,000 people.

OpenAI is strong but they're still dependent on MSFT.

cuckatoo
1 replies
1d2h

MSFT needs companies like OpenAI to give Azure credits to for their valuation to continue soaring. The deferred revenue on their balance sheet from the unspent Azure credits they give as investment are worth much more to their market cap than $80B.

jart
0 replies
1d2h

It sounds like they took the Federal Reserve's business model and applied it to computing.

consumer451
1 replies
1d3h

Just to play this out, what possible moves would OpenAI make at this point that they wouldn't have until this happened?

Altman is out there trying to raise ridiculous sums to get away from Azure, didn't he make the first move here?

bhouston
0 replies
1d2h

I think the main move would be some type of true AGI that leads to a hard takeoff scenario, but it isn't clear we are close to that or not.

Basically something that is more than just another bump in the scorecard for GPT 5 over GPT 4. Otherwise it is still just a horse race between relatively interchangeable GPT engines.

yellow_postit
0 replies
1d2h

OpenAI will want to expand to all clouds to increase TAM.

Microsoft will want to avoid things regulators in the current regime will go after.

This seems like a step towards both and ultimately good for developers as it seems likely to bring costs down by increasing competition.

kibwen
0 replies
1d3h

There are no consequences for Microsoft. It owns a 49% stake in OpenAI, so the only action that OpenAI could take to hurt Microsoft would be to deliberately destroy its own value.

karmasimida
0 replies
11h13m

They are under an exclusive contract with MSFT, they can't do anything.

bookaway
14 replies
1d2h

It's also a good play to try to take resources away from local, self-hosted "Feasible AI" solutions. With compute resources, I think Microsoft hopes Mistral skews their focus and resources towards large models that can run only run in the cloud, trying to lure them away with the bait: "Don't you want to build the best AI possible, independent of compute?"

I'd be surprised if they didn't consider the notion that they are hitting to birds with one stone: OpenAI and Indie AI.

loceng
4 replies
1d

From my understanding, which may be wrong, you only need the massive compute resources initially to create a compiled vector space LLM - and then that LLM once compiled can be run locally?

This is why anti-CSAM measures policy is possible so compiled-release LLMs can have certain vector spaces removed before release; but apparently people are creating cracks for these types of locks?

extr
3 replies
19h13m

You are a little confused. There’s no “compiling” of LLMs. It’s just once it’s trained, inference takes less compute than further training. So you can run things locally that you couldn’t necessarily train locally.

Not sure where you are getting the CSAM bit. We aren’t that good at blanking out weights in any kind of model, certainly not good enough to lobotomize specific types of content.

loceng
1 replies
18h29m

Thanks for the clarification.

The CSAM bit seems to then be propaganda from at least one AI company putting out PR to falsely quell people's concerns about their LLMs being able to generate content involving children that's sexualized.

I've yet to see details of how much compute-minimum server requirements are necessary to run LLMs. Maybe you know a source who's compiling a list in a feature matrix that includes such details?

dwaltrip
0 replies
16h13m

Large LLMs like gpt-3 and gpt-4 need very serious hardware. They have hundreds of billions of parameters (or more) which need to be loaded in memory all at once.

JyB
0 replies
18h47m

I love that you are using the word lobotomize.

dhruvdh
3 replies
1d1h

It's not like Microsoft is working on "Windows AI Studio" [1], or released Orca, or Phi. It's not like there's any talk of AI PCs with mandatory TOPs requirements for Windows 12. Big bad Microsoft coming for your local AI, beware.

[1] https://github.com/microsoft/windows-ai-studio

whimsicalism
1 replies
1d1h

You’re downvoted for the snarky tone I guess, but you’re absolutely right

peteradio
0 replies
1d1h

Even easier to rug pull your own teams project than someone else's.

drakenot
0 replies
15h13m

The whole embrace, extend.. ?

Mistral Remove "Committing to open models" from their website

That was 5 hours ago.

Without having insider details it is hard to know why, but the coincidence of timing with the Microsoft deal is not lost on me. It could have even been a stipulation.

smoldesu
1 replies
1d1h

I don't see why Mistral would acquiesce. Like the other comment says, Microsoft has a lot of chips on the table for local AI. They didn't even mention DirectML, ONNX or Microsoft's other local AI frameworks - suffice to say Microsoft does care about on-device AI.

So... would Mistral deliberately sabotage their low-end models to appease Microsoft's cloud demand? I don't think so. Microsoft probably knows that letting Mistral fall behind would devalue their investment. It makes more sense to bolster the small models to increase demand for the larger ones, at least from where I'm standing.

moneywoes
0 replies
2h51m

what’s the most promising?

teh_infallible
0 replies
1d

So.. Embrace, Extend, Extinguish?

isoprophlex
0 replies
1d

Moving straight from Embrace to Extinguish, why not!

Zuiii
0 replies
14h2m

Antitrust. Using their dominance in one market to destroy another. I hope the EU tares them a new one if the us doesn't.

zitterbewegung
11 replies
1d3h

I totally agree it’s also like the move where Microsoft is at least supporting Linux on their systems and cloud as not a backup but to just close you into their ecosystem . Honestly I could see Microsoft buying Huggingface.

bhouston
3 replies
1d3h

Yes, Microsoft doesn't have to pick the sole winner in AI, but rather they could just start eating the AI ecosystem bit by bit so that they win by default. It is what large players can do. May open themselves up to some scrutiny for too many acquisitions and reducing competition though, but that is a separate issue.

solardev
0 replies
1d2h

"Microsoft recommends OpenAI as your default overlord. Did you know it can do everything your current AI can do, sometimes better, but always more profitably for us? [Switch now] [Ask me again in 30 seconds]"

lostemptations5
0 replies
1d3h

And they are used to that issue too. A long history of it.

datadrivenangel
0 replies
1d2h

This is how microsoft has been doing data for at least 10 years (See databricks).

Step 1: Get the industry leaders to be purchasable via Azure. Step 2: Slowly build your own clone and start stealing user share even though your offering is still worse.

btown
2 replies
1d2h

This has me thinking about the context behind the striking quote in https://www.theinformation.com/articles/how-microsoft-swallo... (May 2023, months before the OpenAI drama):

Nadella [in December 2022] abruptly cut off Lee midsentence, demanding to know how OpenAI had managed to surpass the capabilities of the AI project Microsoft’s 1,500-person research team had been working on for decades. “OpenAI built this with 250 people,” Nadella said, according to Lee, who is executive vice president and head of Microsoft Research. “Why do we have Microsoft Research at all?”

At the same time, even as the company began weaving OpenAI into the fabric of Microsoft’s products, Nadella decided not to abort Microsoft’s own research efforts in AI. During the tense exchange at the December meeting between the Microsoft CEO and Lee, other executives spoke up to defend the work of Microsoft’s researchers, including Mikhail Parakhin, who oversees Microsoft’s Bing search and Edge browser groups, Lee said. After grilling Lee in the meeting, Nadella called him privately, thanking him for the work Microsoft Research had done to understand and implement OpenAI’s work in a way that passed muster for corporate customers. Nadella said he saw Lee’s group as a “secret weapon.”

While this is entirely speculation, it's easy to imagine that there are many levels of PR magic going on here, to share a quote that on the surface feels "leaked" and "explosive" but, among investors and clients who read beyond the (very good) paywall, actually shores up a narrative that Microsoft has a capability that significantly augments OpenAI's, and allows the existence of MSR to become headline news without even needing a product release.

The Mistral deal feels like yet another step in this direction. Microsoft is not afraid of seeming "messy" in the press as long as it can control the narrative around its value-add to customers in the context of its partnerships. By contrast, the rest of FAANG's more consumer-facing positioning makes it a lot harder for them to maneuver in a similar way.

karolist
1 replies
1d2h

Nadella [in December 2022] abruptly cut off Lee midsentence, demanding to know how OpenAI had managed to surpass the capabilities of the AI project Microsoft’s 1,500-person research team had been working on for decades. “OpenAI built this with 250 people,” Nadella said, according to Lee, who is executive vice president and head of Microsoft Research. “Why do we have Microsoft Research at all?”

The answer to that is till Google released the Attention is All You Need paper in 2017 there were no breakthroughs allowing models as we have now to be built, OpenAI being a small and nible team picked up on which direction the wind is blowing with LLMs and quickly brought a product to market whilst MS just did what corps do - move slowly (same for Google etc).

zitterbewegung
0 replies
1d1h

Microsoft research has also been not solely devoted in AI I have seen much in quantum computing and programming language research and general computer science .

JimDabell
2 replies
1d3h

I’ve had similar thoughts, Microsoft buying Huggingface would be very similar to them buying GitHub.

bugglebeetle
1 replies
1d

Please, god, no. I can’t think of two more antithetical companies.

jahsome
0 replies
13h33m

Would you mind elaborating why? I'm not super experienced in the AI world, and barely use Hugging Face. Frankly, the name makes it difficult to take it seriously.

probably_satan
0 replies
1d1h

They did that because they can't compete with Linux and had no relevancy in the tech world outside of providing business users with terrible software

bugglebeetle
6 replies
1d2h

Yes, it’s their tried and true maneuver: embrace, extend, extinguish.

blackoil
5 replies
1d1h

Can we stop having this comment in every Microsoft post. It is like people have no clue what EEE is/was, but if it is MS let's post this.

bugglebeetle
3 replies
1d1h

Uh, sorry, but this seems pretty consistent with trying to co-opt and kill open source AI competition:

[EEE] describe its strategy for entering product categories involving widely used standards, extending those standards with proprietary capabilities, and then using those differences in order to strongly disadvantage its competitors.
ramon156
1 replies
23h33m

Also I've never heard the term myself

jraph
0 replies
13h10m

I suggest reading up on this, it is an important element that significantly shaped the computer world.

But yeah, not sure it is at play here.

jodrellblank
0 replies
13h47m

"The US tech giant will provide the 10-month-old Paris-based company with help in bringing its AI models to market. Microsoft will also take a minor stake in Mistral although the financial details have not been disclosed"

Where are the "widely used standards"? Where are the "extending the standards with proprietary capabilities"? Where is the "strongly disadvantaging competitors"?

knowriju
0 replies
11h11m

We should just automete these comments: MS: Oh No! Embrace Extend Extinguish ! Google: Oh No! killedbygoogle ! Meta: Oh No! So much Ads ! Apple: Oh No! Evil App Store policy !

dkjaudyeqooe
1 replies
1d2h

The pro is that MS buys more AI hype to pump up their share price.

The con is that MS attracts more attention from regulators.

datavirtue
0 replies
16h18m

Those same regulators who check Microsoft's stock price regularly to see how 25% of their retirement plan is doing?

tailspin2019
0 replies
23h57m

This is just a smart move by Satya Nadella

Diversifying their AI bets definitely makes total sense. If this wasn't their strategy originally, it almost certainly became so the moment the OpenAI board fired Sam Altman.

It's easy to make simplistic judgements from the outside, but with the limited information we have, it does seem like Satya Nadella came out of this OpenAI debacle looking pretty competent.

It's hard to reconcile the fact that the Microsoft that handled the unexpected OpenAI issue so well is the same Microsoft that seems intent on literally setting fire to their flagship product! (Windows)

kernal
0 replies
21h46m

A smart move by Microsoft is to not be reliant on another company for their AI needs.

epolanski
0 replies
1d

Putting multiple bets and having multiple partnerships is smart regardless of OpenAI drama.

This way Microsoft is less dependent on a single deal and can diversify their offering based on use cases.

behnamoh
0 replies
1d2h

Satya is known to play 4-D chess. With this deal, MSFT is at least two dimensions ahead of the competition. /jk

nojvek
34 replies
21h44m

Microsoft knows they won big in Cloud via Azure Big as in 100s of billions.

They know someone is going to win big in AI. They lost big on the mobile OS. It is the biggest blunder of Microsoft.

They don’t want to lose on AI. Better to hedge your bets.

They have so much money lying around. Better to invest where they think the puck will be.

Yes AI is hyped but I doubt the bubble will burst. Cheap human like intelligence if achieved via AGI is gonna make them 10x more valuable.

Meta is all in, Google is all in, Microsoft is all in.

Tesla is betting they’ll be the biggest robotics company on the planet.

Getting to AGI is like getting to first nuke for tech behemoths.

dmitrygr
9 replies
21h32m

Getting to AGI is like getting to first nuke for tech behemoths.

Except, unlike AGI, nukes happened. We are no closer to AGI than we were when we lived in caves. We found a small local maximum (LLMs), I have seen NO evidence that there is a path from here to AGI.

fragmede
6 replies
21h25m

Even if LLMs are a local maxima, given that we're even asking that question implies that we are closer to AGI than caveman who didn't have GPUs, even if we don't know exactly how much closer. Unless it takes us 12,000 years to get to AGI, which, it could, since we don't have it yet, but assuming we get AGI before 12,000 years are up, which is when the end of the Paleolithic era was, then yes, we can say we are closer to AGI than cavemen, regardless of if LLM or GPUs are how it comes about. If it never does, then this comment will age poorly in 12,000 years, but I'm okay with that.

dmitrygr
5 replies
21h18m

given that we're even asking that question implies that we are closer to AGI than caveman

"Am I close to having gills and breathing underwater?" The fact that I asked, changes nothing.

joelfried
4 replies
20h53m

You don't think having functional computers being capable of trillions of operations per second is even the slightest bit of an improvement over use of literal stones?

I am genuinely curious: What do you imagine would be the kind of thing that would be a meaningful step towards AGI?

I have no idea when we'll get there for real, but it seems a pretty big assertion that nothing invented in the last 150 years even helped. So what do you think would help?

dmitrygr
3 replies
20h19m

Ok. I’ll rephrase. In caves we were 0% there. Now we’re 0.00001% there :). Maybe. May need a few more zeroes. What we have helps probably, but hard to be sure how much. Maybe we need 1e40 FLOPS for AGI. In that case going from 1FLOP (human in a cave) to 1e14 FLOPS (a GPU now) is irrelevant.

I’ll believe we’re closer when we have a computer solve a novel problem that is not a simple pattern match to a similar solution. When a computer can reply to “write me a JIT translator from ARMv5 to ARMv7M” with working code. That takes actual thought and we’re not even close.

dmitrygr
1 replies
18h30m

That covers 0.001% of the work - the part that my polit-sci-major no-programming experience sister just explained to me given the prompt. Now do the hard parts :)

fragmede
0 replies
9h33m

The hard parts are just an implementation detail. Let's say a human took 15 minutes to draft all that up. It's about 800 words, and given a typing rate of 50 wpm, it would take that long to just type that in, so that seems reasonable. If that's only 0.001% of the work, it means it would take 3 years to finish the project.

3 years to write a JIT from ARMv5 to ARMv7M?

I think your numbers might be off :)

btw, your polit-sci-major no-programming experience sister seems pretty clever if she was able to understand all that. I ran it past a few cavemen I came across and they didn't quite understand the nuances surrounding runtime environment management.

thatoneguy
0 replies
21h17m

I think the real nuke will be that we're already at some basic level of AGI because human consciousness isn't as complicated as we think it is. I'm guessing we're only 10 years out from lowly generative AI getting advanced enough to effectively be what think of AGI now.

edouard-harris
0 replies
20h57m

This seems like a remarkable chain of reasoning. Out of curiosity, what would convince you that there is a path from LLMs (say) to AGI? What would you expect progress along that path to look like?

elorant
8 replies
21h4m

It’s not just that. Microsoft is in a better position today to make big money from AI because their products are in millions of corporations. B2B is their bread and butter. Google isn’t in that position, neither is Apple, Meta or most AI companies out there. Microsoft has built trust over decades of selling OSes and offering legendary backwards compatibility in anything they build. This is what the big spenders want.

karamanolev
3 replies
21h1m

How is Google not? A quick search points that Google Workspace has a huge market penetration.

kccqzy
0 replies
19h46m

The Google Workspace is roughly comparable with Office 365. But Microsoft has a lot more than just Office.

elorant
0 replies
20h55m

Those are all web toys. We’re talking about databases and CRMs and programming languages and all the stuff that make the backbone of big corporations.

SkyPuncher
0 replies
20h57m

Correct, but there’s typically a divide between workspace tools and infrastructure for ownership, compliance, and approval internally.

In a prior life, our IT manager was the owner of Microsoft productivity products and I was the owner of Azure. We both had drastically different risk profiles and governance needs.

jjtheblunt
3 replies
20h40m

i think you're overlooking iOS instances with "neural" accelerators in the pockets of the people in those corporations.

elorant
2 replies
20h30m

And how will that materialize? Will Apple build a new AI chip that would be affordable enough to embed in every iPhone, or whatever new device they will invent? Or will it all be cloud based where it should be able to support hundreds of millions of users?

Microsoft already has the cloud infrastructure ready. They don't need to build a new device, or a new operating system, or whatever. They're milking the AI cow as we speak.

jjtheblunt
1 replies
19h38m

Apple has had "AI chips" in the iOS devices for years already. I wonder what you define as an AI chip, though, since we could be discussing different sorts of custom accelerators.

elorant
0 replies
19h25m

Look at what a company called Groq does. Their chip can generate 500 tokens/second from a model like Llama 2 70b.

https://groq.com/

MuffinFlavored
6 replies
21h40m

They lost big on the mobile OS. It is the biggest blunder of Microsoft.

I've watched Bill Gates tell Andrew Ross Sorkin during an interview "if it wasn't for the FTC investigating us for the anti-trust lawsuit, we wouldn't have been distracted/took our eye off the prize on mobile"

How sure are we Microsoft "has what it takes/had what it took" to deliver a phone + operating system as polished as Apple?

Windows 11 is very much measurably worse than Mac OS Sonoma. Littered with ads, in between old UI + new UI patterns, etc.

That's 2024

I'm not super confident I'd prefer a Microsoft/Windows mobile OS and therefore I'm not super confident they could actually have delivered a good one

mattmaroon
1 replies
20h59m

I think Windows Mobile could have been better than Android. In fact, at one point it debatably was. Personally I think Windows is considerably more usable than OSX (yeah I know that’s going to offend people and raise a near-religious argument) and while I won’t say Microsoft’s hardware is or ever was as good as Apple’s, I think they are far better at UI/UX and hardware than Google. Actually I think they’re better at almost everything than Google.

They could have easily been where Android is now if they were two years faster. I think we’d just have a different duopoly and iPhone would still be basically where it is today.

blagie
0 replies
20h23m

I won’t say Microsoft’s hardware is or ever was as good as Apple’s, I think they are far better at UI/UX and hardware than Google.

Microsoft-proper hardware is crap, even compared to high-end Google. However, that doesn't matter. My Thinkpad eats Apple hardware for breakfast (as well as anything from Google or Microsoft). Dell has a few nicer laptops in the Precision line which do almost as well as Thinkpad, and definitely better than Apple / Google / Microsoft.

It's the ecosystem.

Actually I think they’re better at almost everything than Google.

I'd agree, except:

- Hardware

- Office 365 typing synchronization

Google Pixel Pro is quite good (as are most top-of-the-line Google phones), as was the Chromebook Pixel (their top-of-the-line Chromebook, when they made it).

quantisan
0 replies
21h4m

I had multiple Microsoft Lumia as my primary driver for a couple years. I've been on iPhones for 10+ years total. There were some good (better user interface) and bad (Live logins required in random places) with Windows Mobile. Like Windows, they have pockets of good ideas fractured by their lack of cohesion.

fragmede
0 replies
21h30m

It doesn't matter if it's any good, it just has to be functional, which Microsoft proved they can do, since windows phones existed. Microsoft has the business side (aka money and patience that Google doesn't) to make it happen. Look at how well Microsoft Teams is doing. Look at Bing, now that it's got ChatGPT behind it.

blagie
0 replies
20h29m

Microsoft has gotten pretty good at software development. Satya Nadella has been doing a surprisingly good job reforming everything about the company.

That doesn't mean Windows wasn't absolute dogs--t for many years and isn't riddled with a pile of technical debt. Apple, in contrast, started with NeXTStep. I agree about the specifics you mention (built-in ads and spyware), but the core problems with Windows predate modern Microsoft, and can't really big fixed.

If, in 2024, Microsoft decided to invest in building a real mobile OS, I think they could do okay. The bigger problem would be lack of app ecosystem, and the chicken-and-egg problem with that and users. It's not clear even the best mobile OS could displace Android + iOS.

If I were Microsoft, and I wanted to get in, I'd probably fork Android, to maintain app compatibility. Doing that well would mean killing the goose which is currently laying the golden eggs, though, as it would require replacing Google apps with Microsoft ones. A lot of what makes Android work are free Google accounts, whereas Office 365 is $100 per year. I don't think a Microsoft phone would be competitive unless it had all the same stuff for free, and likely more.

I actually feel like Android is starting to be a little bit vulnerable; a privacy-preserving, non-user-hostile version could have pretty good uptake. Again, not the business Microsoft is in.

AlienRobot
0 replies
20h43m

I find this ironic because you could say many of the problems with modern Windows (& Linux) GUI stem from distancing the design from the traditional desktop paradigm and seeking to mobile-fy everything, specially with large click targets.

michaelbuckbee
2 replies
21h33m

Microsoft's gunning for a far more short-term win: Search.

Bing has been an also-ran in the search world for a decade (single digit % of search volume compared to Google's 90%+), and AI has shown the first real crack in Google's monopoly.

The future is pretty clearly going to be directly getting AI answers to questions and much less looking at a page of 10 blue links and it's still really unclear if Google is going to make the transition.

chasd00
1 replies
20h9m

The future is pretty clearly going to be directly getting AI answers to questions

with a subscription model, no ads, and therefore no seo manipulation of content/answers. I'm all for it!

nimbius
1 replies
21h15m

Microsoft knows they won big in Cloud via Azure Big as in 100s of billions.

Malarkey. Microsoft got where they are in cloud by dogfooding every single piece of software they wrote into their cloud platform, every single service they ran on their gaming platform, and every single user on github and minecraft. After that they turned to corporate customers and forced them into the cloud as well and finally made every single end user of windows sign up too (cant miss any of those precious KPIs.) If they won anything it wasnt through genuine consumer desire to use Azure. The efforts were mostly a shuffling of deck chairs and pump-job similar to the IIS wars on netcraft back in the day where M$ would pump their IIS numbers with static content served from parked websites at hosting providers they paid to switch from Apache.

They have so much money lying around. Better to invest where they think the puck will be.

they also have a track record of building things no one wants and ruining things everyone likes. Minecraft and Github are demonstrably worse in a lot of ways than they were before redmond took the helm. having a lot of money doesnt make you clairvoyant.

Meta is all in, Google is all in, Microsoft is all in.

who the fuck cares? these are all companies that exist in the nadier of their innovation. that they collectively scrape barrel to come up with tech memes for the business kids isnt exactly remarkable outside the fact they havent delivered anything of value in so long its surprising we still have to hear about them at all.

Tesla is betting they’ll be the biggest robotics company on the planet.

The guys who cant get self-drive right? run by the same guy who pedaled twitter into the ground? sure.

AI is a meme for these companies...a parlour trick they use on the money pump for just another year longer before its reinvented into some other nonsensical sci fi pablum for general consumption under the late hour of capitalism.

ciberado
0 replies
20h52m

Confrontational replies to comments that doesn't provide actual data or concrete examples to support the counterarguments are not very useful. It would be interesting to know why do you find in GitHub worse than before, or why Minecraft is ruined.

Edit: typo fix.

Jackson__
1 replies
21h29m

And yet, it's kinda looking like everyone except OAI has hit the same pre-gpt4 capability wall.

I'll hedge my bets on whether AGI is even possible entirely on how much of an improvement we'll see with GPT5. If it is just a marginal improvement, that's basically it for the current ML bubble.

chasd00
0 replies
20h7m

that's an interesting take, i wonder what the graph looks like of model capability over time. If GPT5 doesn't fit in the curve and a year goes by with no one else showing the same rate of progress... is that it for the LLM hype machine?

Edit: i was thinking while walking my dogs, if GPT5 was another great advancement and especially if AGI was around the corner then why would Karpathy leave OpenAI?

netdur
0 replies
21h3m

I believe Google stands apart; although their current offerings sucks, they possess all the necessary components, including software, data, and hardware.

neel8986
32 replies
1d3h

At this point OpenAI should definitely think of some sort of partership with AWS and GCP. Otherwise they will be just one of the few models

bhouston
18 replies
1d3h

I personally didn't realize how fast other models would catch up to OpenAI.

There is a whole set of models now (and some like Meta are purposely trying to undermine OpenAI competitive advantage via open source models) and they are relatively interchangeable with nearly no lock-in.

OpenAI's main advantage is being first to market and having the strongest model (GPT 4), and maybe they can continue to run ahead faster than everyone else, but pure technical leadership is hard to maintain, especially with so many competitors entering.

behnamoh
8 replies
1d2h

I personally didn't realize how fast other models would catch up to OpenAI.

They haven't though. Gemini is vaporware and other models are not as good as GPT-4.

bhouston
3 replies
1d2h

Mistral Large is quite close to GPT 4 per the stats from today:

https://mistral.ai/news/mistral-large/

Now once OpenAI launches GPT 5 well I am sure other models won't look so good, but right now these other models are approaching GPT 4 capabilities.

parineum
2 replies
1d1h

I'm pretty bearish on GPT 5 being better than 4. With how neutered 4 has gotten over time, I'd be surprised if GPT 5 is able to perform better with all the same restrictions that GPT 4 has. GPT 4 is less and less willing to actually accomplish a task for you than it is to tell you how you can do it. It looks more and more like Markov chains every day.

staticman2
1 replies
1d1h

Did you mean bearish?

parineum
0 replies
1d

Yes I did, thanks.

donny2018
2 replies
1d1h

For now. As others have said, there is no technological “moat” in this business that could prevent others from catching up.

Perhaps the best way for Open AI is to become THE established AI services company. AWS is still the leader in cloud computing space, and only has Azure competing, despite the fact that other big companies are also technologically capable of building similar products.

nicce
1 replies
1d1h

AWS is still the leader in cloud computing space, and only has Azure competing, despite the fact that other big companies are also technologically capable of building similar products.

What happened to GCP? I personally switched away because of the bad experiences.. but is that happening in scale as well. I see it barely mentioned these days.

kccqzy
0 replies
19h48m

GCP is suffering from bad sales strategy and lack of existing enterprise relationships.

PunchTornado
0 replies
1d2h

I find gemini advanced better than gpt4. it is my go to option now.

Marlinski
7 replies
1d2h

Their main advantage for now is their super clean API. Open source alternative are already on par with GPT-3.5 and 4 capabilities, they just don't have as good a package but that could change rather quickly too.

tombert
3 replies
1d

Is that true? I was running Llamas on my laptop a few days ago, and it was giving measurably worse results than ChatGPT. I think it was the uncensored 13B model, but if you got something that's on par with ChatGPT that I can run on my own hardware I'm pretty interested.

bhouston
2 replies
22h35m

13B models probably cannot directly compare with ChatGPT 4 which maybe +1T parameters or a 5 way MoE of 200B each - or something like that. So you can not likely run a model competitive with ChatGPT locally in the near term.

tombert
1 replies
22h11m

I have a server with a bunch of PCIe slots and like 4 Nvidia GPUs with 24GB of RAM each. What's the best model I can realistically run?

koito17
0 replies
23h27m

Open source alternative are already on par with GPT-3.5 and 4 capabilities

I'm not sure if this is true. With GPT-4, I can successfully ask questions in Japanese and receive responses in (mostly natural) Japanese. I have also found GPT-4 capable of understanding the semantics of prompts with Japanese and English phrases interleaved.

Out of curiosity, I tried doing the same with local models like Mistral 7B and I could never get the model to emit anything other than English. Maybe it's a difference in training data, but even then, GPT-4 has an allegedly small set of training data for non-European languages.

generalizations
0 replies
1d

What open source alternative is on par with GPT4?

fakedang
0 replies
1d2h

Mistral's API was designed to be practically interchangeable with the OpenAI API.

moffkalast
0 replies
23h2m

Sure, but it is somewhat disheartening to see GPT 4 still being the king by a clear margin after a full year, especially since it's been nerfed continuously for speed and cost effectiveness.

hef19898
11 replies
1d3h

Honest question from someone who never touched one of those models, is OpenAI anything else than the first and most hyped model developer?

IanCal
8 replies
1d3h

IME nothing is as good out of the box as gpt4 for many tasks.

JackFr
6 replies
1d3h

Is gpt4 as good in non-English uses? It's not clear to me that it would be particularly important or advantageous, but does Mistral being based in Europe and polyglot first make it interesting vs. gpt4 in some dimension?

whimsicalism
1 replies
1d1h

gpt4 is going to better than other models in every language except maybe chinese

staticman2
0 replies
1d1h

Depends on the task. Gpt 4 isn't necessarily better at translation than Claude. I'm aware of no benchmarks on this.

moffkalast
0 replies
22h59m

It's the best multilingual model out there and it's not even close.

Especially in terms of open models Mistral's are the most multilingual but outside a few handpicked ones the level of proficiency is just too poor for any real usage.

barrell
0 replies
1d1h

In my experience it’s not such a simple question. If you want to be able to speak in nuanced non-English and have it pick up on the intricacies, or have it respond in rich correct non-English, then it’s not the best model (Cohere recently released an aya model that I would recommend checking out if this is your use case).

If you want to be able to give basic commands and have the model reason about the logic behind your commands, gpt 4 is still the best, even in minority languages.

TwentyPosts
0 replies
23h44m

From what I heard Gemini might be better at translation tasks than GPT4.

Al-Khwarizmi
0 replies
1d2h

I guess it might depend on language, but as a Spanish speaker who sometimes uses LLMs in Spanish, I'd say the gap between GPT-4 and most of the competition (Mistral included) is actually larger in Spanish than in English.

wing-_-nuts
0 replies
22h43m

I don't disagree with you, but an open source model fine tuned for your use case, and embedded with your data is probably going to be way better at many companies uses cases than GPT4 is.

moralestapia
0 replies
1d2h

Yes, GPT-4 still rules, downside is it's expensive and relatively slow.

amf12
0 replies
15h21m

I'd say you should compare the models for your use case. Which is better depends on how much you're willing to pay, what kind of problems you need help with, speed, ease of use.

changoplatanero
0 replies
1d2h

Their problem is that they had to agree on an exclusive relationship with azure in order to secure funding from Microsoft

reissbaker
20 replies
19h29m

Huh, I wonder if this comes with an OpenAI-like exclusivity deal for Mistral's closed models. If so, Satya just wrapped up pretty much the two largest names in AI other than Google.

Sad but expected that Mistral is releasing its newer models as closed source. There was never a clear revenue stream from the open models.

I'm confused why they aren't allowing their models to be finetuned, though — that seems like an obvious way to compete with OpenAI, which only allows finetuning of its pretty weak gpt-3.5-turbo. Sure, Mistral can't quite beat GPT-4 yet, even if it's close — but allowing finetuning would likely result in way better than GPT-4 performance on a broad variety of tasks (as long as you picked a few to specialize in).

IAmNotACellist
16 replies
18h59m

other than Google

Which has proven worse than useless. It's a brand-damaging laughing stock.

JeremyNT
11 replies
15h42m

I've used Gemini and find it no worse than ChatGPT for the kind of questions I ask of it.

tim333
1 replies
6h39m

By the way the links didn't work, for me at any rate.

themoonisachees
1 replies
11h40m

Pretty sure gpt-4 is doing image-to-llm directly while Gemini also runs the image through Google lens which is tailor-made for OCR. I think that's a good thing to create a better product overall but the performance isn't really comparable

kingsleyopara
0 replies
10h2m

At least in my tests, GPT-4 also seems suspiciously good at reading text so it’d be surprised if they weren’t also doing this.

jonny_eh
3 replies
14h51m

Complaining about Google's AI has become some kind of right-wing culture war thing.

AnarchismIsCool
2 replies
13h57m

Very-leftist here. They're becoming an everyone culture war thing because what they did was sufficiently brain-dead as to make everyone look bad. It was sufficiently skewed as to create a number of extremely racist images just because it was trying way too hard to avoid it.

That said, those of us in-the-know understand that this is just part of the cycle and they'll go back to being annoyingly bland and completely devoid of entertainment value soon.

tim333
1 replies
6h35m

Most people are attributing this to woke culture at google but it's possible it's just a cock up. I mean someone told it to use diverse images and forgot about historical plausibility.

JeremyNT
0 replies
3h43m

This is an early product, and monkeying with the system prompts to try and prevent some classes of offensive results is sure to have the kind of unintended consequences seen with Gemini's image generation.

You can learn a lot about people's media diets based on how they perceive these issues with Gemini. Those who quickly proclaim that the product is a shameful embarrassment are almost universally very online, particularly in right wing spaces like twitter/x, even if they wouldn't consider themselves to be "anti-woke" personally.

Taking a step back, Google is clearly going to mess with the system prompt over time to correct this stuff, and bystanders' degree of fixation on this specific issue tends to suggest a certain set of politics. People have bemoaned how "woke" ChatGPT was in the past too, and OpenAI has spent a lot more time under public scrutiny iterating on their own system prompt.

forgotmypw17
0 replies
7h24m

I've found it much better for anything that needs book knowledge.

ThinkBeat
0 replies
5h23m

Since I am on the free tier of both I guess I have only used GPT 3.5 and whatever the free offering of Gemini is.

I find Gemini to be significantly slower. I am sure Google can fix that on premium tiers.

I also found Gemini to be less able. It has fewer tasks / fields it will answer about. Gemini is more honest though.

ChatGPT seems to nearly in all cases say "Sure I can do that" and then return some non-coherent answer.

marginalia_nu
3 replies
17h48m

Google's always been kinda garbage at marketing. Their challenges don't seem to be technical, as much as organizational.

nine_zeros
2 replies
17h33m

It is organizational. The amount of time people spend on performance reviews and other organizational dances can easily be harvested for more productive activities.

But no. Thousands of employees gotta do the dance for blind committees.

knowriju
1 replies
11h14m

Wow, somebody has a deep wound that still hurts.

herewego
0 replies
2h20m

Or they’re simply commenting on Google suffering from a commonplace malady seen in many enterprises.

behnamoh
1 replies
18h21m

I'm confused why they aren't allowing their models to be finetuned

They already have difficulty serving the large model (as evident in their pricing). Finetuning would introduce even more compute challenge.

reissbaker
0 replies
14h46m

Sure, but why not let Azure offer it as a service?

option
0 replies
18h23m

It's not close. Long tail is probably as hard for LLMs as it is for search.

finnjohnsen2
15 replies
1d

It is the easiest to assume the worst for sure, but Microsoft is not the same as the Bill Gates era so Im gonna be a cautious optimist on this one. Lets hope it is to promote Azure, and they dont push the OpenAI route when it comes too openess. Which is closed and a big loss for the world and a disappointment

tombert
7 replies
23h52m

They're not the same, but they're still a for-profit company.

90's era Microsoft wasn't evil for the sake of being evil; they were evil because they felt that monopolistic practices were the easiest way to increase their share price. They have a responsibility to their shareholders to try and maximize their share price and so it's hardly unsurprising that they did the infamous Embrace Extend Extinguish, and until regulators stepped in, such practices worked pretty well.

Most companies don't get large enough to form any kind of real monopoly, so it's easy to get on a high-horse. It's also easy to act like it was just a product of "those people", but I fundamentally think that it's a natural consequence of a company that has achieved nearly-total market dominance.

I have very little faith that a multi-trillion-dollar company is going to prioritize what's best for the world. Fundamentally, I think that if they feel they can get away with it, they'll revert to monopolistic tendencies and try and increase share price.

I'm not just picking on Microsoft here either; replace them with basically any other near-monopoly in tech and my criticisms still hold.

finnjohnsen2
5 replies
23h23m

90's era Microsoft wasn't evil for the sake of being evil

You are to forgiving. I don't care too much about the why behind the evil in the software business and this is probably why my rants fall under Godwin's law too often. I won't this time.

They have a responsibility to their shareholders to try and maximize their share price

... but I was damn close.

I agree with your whole point about Microsoft. But I don't think it's the same company anymore, I like some of the recent stuff, and I trust their lack of monopoly on the web. For now.

exe34
4 replies
20h10m

You are to forgiving

I think the point is that companies aren't good or evil, they are amoral non-aligned super intelligences that maximise shareholder profits.

Loughla
3 replies
15h45m

Is it the intent or outcome that is the evil?

If I am intent on killing a million people, but can't, but a company kills a million people via inadvertent faulty products, which is evil?

leplen
2 replies
15h12m

If a hurricane kills a thousand people is it evil? It's not really clear whether we're better off trying to treat corporations as moral actors and constantly being surprised when they aren't, or whether we should treat them as amoral quasi-natural processes no one has full control over and construct regulations around them the same way we do for earthquakes.

patcon
0 replies
12h29m

I like the metaphor, but... a hurricane doesn't write values documents or employee handbooks. These things do inform actions in many orgs, and constrain opportunities to a degree.

I disagree with this larger idea (that others here are implying more indirectly) that companies are all the same, and like some force of nature. They are guided and there are better and worse ones.

Having said all that, I'm still an anti-capitalist to the extent that one can be one ;)

jraph
0 replies
13h5m

We should still treat corporations as moral actor (without being naive of course) because they need to be accountable for what they do. This is why you can sue a company. You can't sue an earthquake.

Some companies try to do the right things too. Bad corporate behavior should not be normalized this much.

bossyTeacher
0 replies
19h2m

I have very little faith that a multi-trillion-dollar company is going to prioritize what's best for the world

You don't have to have faith. Why would anyone EVER believe that? By definition, a company is just a profit making machine. People believing that making rich people richer will necessarily make the world a better place are living in a koolaid-boosted fantasy world

advael
2 replies
15h10m

Everyone who has ever bet against Microsoft trying to enclose or eliminate something it takes an interest in has been wrong. Why should this be any different?

jodrellblank
1 replies
13h41m

Are you using Windows Phone running on Nokia to Bing! and decide, email with Hotmail, chat with Skype, updating your Silverlight blog with LiveWriter, navigating with Autoroute? ... nobody else is, the people who bet against all those things were right.

advael
0 replies
6h43m

So wait, you're trying to interpret me saying "Microsoft tries to close up or eliminate anything it takes an interest in" to mean "Microsoft never loses"? Very different things.

You can't trust Microsoft to act in the interest of open-source projects, transparency in general, or the users of any of the things it buys. Not all of their attempts to harm ecosystems or products work, and this has very little to do with whether they win competitions with other tech giants. In fact, sometimes user distrust stemming from their long history of Embrace Extend Extinguish and other user-hostilities has played into their failures. But like most shambling behemoths of companies, their deep pockets allow them to stay the course through many failures

joe_the_user
1 replies
23h51m

The ads I see on Windows desktop machines (and shenanigans with nonstandard Html extensions and browser defaults and etc...) tell me Microsoft is just as eager to leverage it's monopoly status as ever.

No doubt the company is cautious about some things now but even in these, it will push the boundaries.

huytersd
0 replies
20h18m

Those ads on the login screen are so cheap, wonder why they would do that.

downrightmike
0 replies
19h31m

MSFT just doesn't say the quiet part out loud these days

BirbSingularity
0 replies
23h11m

They already have removed their commitment to open-source. It's not looking good.

blitzo
15 replies
1d3h

I bet sooner or later Microsoft will be in hot water again with antitrust and this time with generative ai

crazygringo
9 replies
1d2h

You're going to have to give some specific reason why though.

Why this would have anything to do with antitrust is not at all obvious to me. Especially when Google has been inventing and acquiring its own generative AI technology that it is competing with.

whimsicalism
6 replies
1d1h

The OAI stake/deal is already under regulatory review and generally EU is perceived as blocking most large tech mergers since the iRobot intervention.

I suspect we are going to soon see political backlash against regulation in the EU as it is becoming very clear that this is causal to their bad capital markets.

nicce
3 replies
1d1h

I suspect we are going to soon see political backlash against regulation in the EU as it is becoming very clear that this is causal to their bad capital markets.

Who would have thought that human rights are bad for business..

whimsicalism
2 replies
1d

What human right do I have to Amazon not owning a vacuum company?

nicce
1 replies
1d

Nothing. But you probably don't see one located in Europe, because they would need to allow strikes, there is good level of minimum wage protection in general and strong privacy laws. It is harder to stalk the toilet breaks for employees.

blackoil
1 replies
1d1h

OAI merger?

whimsicalism
0 replies
1d1h

sorry mis-spoke - OAI stake.

dandy23
0 replies
1d2h

Because MS is everywhere.

croes
0 replies
1d

Tight integration of AI into the OS and Office.

What if users want to user other AI services?

It's Internet Explorer and Media Player all over again.

rapsey
0 replies
1d2h

FTC is pretty useless lately.

krautt
0 replies
1d3h

that's a good problem to aspire toward.

hu3
0 replies
18h14m

Could be worse.

Could be Apple wallet garden charging 30% of any profit their AI API generated.

Jokes aside I wonder why Apple doesn't invest in these prominent AI startups.

It's not lack of cash for sure.

fzzzy
0 replies
1d2h

That's a problem that companies worry about after they have it.

SeanAnderson
0 replies
1d2h

How come?

yogorenapan
8 replies
1d

It’s concerning that the new model isn’t on Huggingface. I hope Microsoft doesn’t stop it from open sourcing the models and code.

unnouinceput
6 replies
1d

Good luck with that wish bro. Microsoft is all about money and there is no money in open source for them.

MH15
5 replies
22h48m

VSCode? Typescript?

sublimefire
3 replies
21h15m

And how exactly this earns money for them? embrace extend extinguish

adrr
2 replies
18h59m

c# is open source. There are other c# implementations like mono which are used in projects like unity.

hx8
1 replies
14h55m

Microsoft SQL Server is closed source. They take open sourcing in a case by case basis, and the primary motivator seems to be money.

jug
0 replies
6h5m

Yes, it's not for altruistic reasons but to lower the barrier of entry.

jacooper
0 replies
19h25m

Vscode is a bad example, almost all useful Plugins are closed source on purpose, the package they ship also includes modifications that aren't open either.

BirbSingularity
0 replies
23h10m

Mistral has already scrubbed their commitment to open source from their website, doesn't look good.

jacooper
1 replies
19h32m

A far better experience than the crappy archive links

Jackson__
0 replies
19h24m

Don't get used to it being around. Google is probably killing them off soon:

@searchliaison

is the cache link in the search results gone forever?

Hey, catching up. Yes, it's been removed. I know, it's sad. I'm sad too. It's one of our oldest features. But it was meant for helping people access pages when way back, you often couldn't depend on a page loading. These days, things have greatly improved. So, it was decided to retire it.

Personally, I hope that maybe we'll add links to @internetarchive from where we had the cache link before, within About This Result. It's such an amazing resource. For the information literacy goal of About The Result, I think it would also be a nice fit -- allowing people to easily see how a page changed over time. No promises. We have to talk to them, see how it all might go -- involves people well beyond me. But I think it would be nice all around.

As a reminder, anyone with a Search Console account can use URL Inspector to see what our crawler saw looking at their own page: https://support.google.com/webmasters/answer/9012289

You're going to see cache: go away in the near future, too. But wait, I hear you ask, what about noarchive? We'll still respect that; no need to mess with it. Plus, others beyond us use it.

From: https://twitter.com/searchliaison/status/1753156161509916873

Jorge1o1
0 replies
18h43m

Archive.is times out if you're behind a VPN or on a VDI.

For example, my work virtual machine on Azure can't ever use it.

ShamelessC
2 replies
1d1h

Anyone else having issues loading that link?

nyoomboom
0 replies
1d

Archive itself has issues for me

jakub_g
0 replies
1d

archive.is is known not to work with Cloudflare DNS (google it for details).

azinman2
7 replies
20h27m

What I don’t get is that mistral is really only possible because of Meta.. and Meta has a user limit on the license. Does that not apply to Microsoft hosting this? Isn’t the economics quite weird here? I’d be pissed if I were meta that someone took my model that I spent millions on and now will host a fine tuned version for money facilitated by Microsoft.

allanrbo
6 replies
19h40m

Why do you say Mistral only possible because of Meta?

azinman2
5 replies
19h15m

It’s built on top of llama 2.

sdrg822
2 replies
19h5m

It is not.

azinman2
1 replies
18h45m

Looks like I’m wrong! I swore this was the case (so many models are), but not this one. Good for them!

Klaus23
0 replies
18h2m

There is a leaked prototype based on llama, but the rest are their own models.

piwi
0 replies
19h7m

I'm not so sure. Do you have a source?

Zambyte
0 replies
19h9m

You got a source for that?

stall84
4 replies
1d2h

Hey when articles are paywalled like this.. What do most of you guys/gals do ? Just comment on the headline?

weberer
0 replies
1d2h

I automatically hit Ctrl+F and search for any comments mentioning "archive". Someone usually posts one within 5 minutes.

euniceee3
0 replies
1d1h

Assume the content is trash and then read the comments for the recap. Like anything that is relevant will get quoted.

skadamat
3 replies
1d1h

Brilliant move by Microsoft. You want to be the one selling shovels during a gold rush, as they say?

whimsicalism
2 replies
1d1h

nvidia is the shovel seller

wing-_-nuts
1 replies
1d1h

quibble, but nvidia is the shovel maker. LLMs require a lot of vram but the underlying hardware requirements are actually fairly simple. There are already efforts in place to create inference specific ASICs.

That'll absolutely eat into Nvidia's profit margins.

skadamat
0 replies
1d

Yeah computation has more layers and sellers, it's not a perfect analogy but good enough I feel

whimsicalism
0 replies
1d1h

American partnerships with EU companies are subject to a lot more regulatory scrutiny in the EU.

izolate
0 replies
1d1h

Not trying to be dismissive of Mistral, but I bet that's a large driving force behind the effort. Usually I'd prefer to focus on the technical aspects, but with the undeniable geopolitical impact of technology/AI, I think it necessitates a discussion.

moralestapia
1 replies
1d2h

Not a dupe, that one is about the LLM release, this one is about the MS-Mistral deal.

ChrisArchitect
0 replies
1d1h

Not only does the official Mistral post have a whole section on the deal, it links out further to MS's official post on the deal also. It's all the same discussion

knowriju
0 replies
11h5m

I don't quite understand the other comments. The press release says Mistra Large is available first on Azure. Which means it will be available on other platforms as well ?

Also, MS's relationship with OpenAI is that they have access to OpenAI's model for use in their products as well as potentially the source code. Is there anything similar with Mistral ? I can't find any such wording anywhere.

mark_l_watson
1 replies
22h55m

With mixture of experts (MoE) models, I would think that Mistral might make a lot of money reselling specific models to big players like Microsoft and Google to augment their MoE systems.

I mostly use either mistral7b or mixtral8-7b for most things, and experiment with other models on the side. In what world will LLMs not become a commodity?

You might answer that question by saying that on player achieves AGI and captures the market, but I think there will be more to AGI than LLMs.

comeonbro
0 replies
22h14m

With mixture of experts (MoE) models, I would think that Mistral might make a lot of money reselling specific models to big players like Microsoft and Google to augment their MoE systems

I am curious about what you mean by this. There is a (very understandable) common misconception about what "mixture of experts" means in current practice.

Mixtral is not a mixture of fully-separable domain experts, like one is an expert at programming and one is an expert at arts and literature. The "experts" are per-layer (and as far as I know, the subject of their "expertise" not even interpretable at this point).

htrp
1 replies
1d3h

Interesting that microsoft is hedging their bets across the foundational model spectrum

Microsoft will also take a minor stake in the 10-month-old Paris-based company, although the financial details have not been disclosed.
KuriousCat
0 replies
1d3h

Is there an alternative for them?

neutralino1
0 replies
1d2h

Satya is buying popcorn.

moralestapia
0 replies
1d2h

Smart move by Satya, to fund both sides of the war.

leumon
0 replies
17h23m

Funny that they say "Its models are “open source”, meaning technical details will be released publicly.". In fact, they aren't only releasing the technical details, they are also releasing the model weights themselves.

jonplackett
0 replies
1d

Clever move.

Thinking about it, Google or Apple should have got in there with Mistral.

hscontinuity
0 replies
22h42m

I'm lost here. Can anyone explain why the speculation of risk resounding around a 'bubble' of LLM and other emerging technologies (TinyLM, VLM, ALM, etc) is rounding a bubble?

Meaning, the dot-com bubble showed a misappropriation of revenue by leading technology companies alongside the capital investorship, mainly VC's and private equity. Basically, everyone spent without producing a product; which lead to the mergers still the basis for many of the largest corporations today (and subsequently the 2008 crash, which further merged companies in the aftermath).

So, for me the difference is the money being spent here in this 'bubble' isn't like parking new mall shops on various websites, then spending the capital to entice buyers to purchase (either for store products or the store itself) - rather - it seems to me that both private and public sentiment is that the future is bright with the right application of these tools, in the right systems, to promote applicable use cases for things we (humans) can leave up to code/machines.

For instance, one of the most impressive applications of TinyLM uses is setting up several (or dozens) of tiny sensors in say a greenhouse. Each sensor can be linked to a central data repository for active monitoring to deploy active controls - be it barometric pressure, moisture content (soil, air), etc. Linking a bunch of these devices and letting it automatically run, along with linking this sensors as the data-providers to machines that may in turn plant, cull, trim, essentially a fully automated greenhouse.

I'm not greenthumb and I lack in-depth knowledge around how greenhouses (modern or old) work. However, it's without understanding the details I do understand it still takes a (or more) human to maintain the greenhouse.

Consider making these kinds of greenhouses (as is standard in some places now with vertical in-door growing farms) completely autonomous.

That's pretty technologically feasible it would seem with the newest applications of (quote) AI. And it's likely very profitable as well.

If I were to take the same ideology and apply to other industries, I find several applications like the one I mention that would absolutely change how we (humans) live in this (soon to be, IMHO) post-human industrialized world.

So where is all of the doom and gloom from - monopoly, errant comprehension of the technologies, or simply dogma?

cuckatoo
0 replies
1d2h

This is just another play to give out more Azure credits to anyone that can feasibly consume them. Azure credits show up as unearned revenue on their SEC filings where they state that they "expect to recognize approximately 45% of this revenue over the next 12 months and the remainder thereafter".

It's wild that you can give out gift cards that make your company's value go up so much more than the gift cards could ever cost you. It's almost like one of those financial schemes that end badly.

boiler_up800
0 replies
1d2h

Just from experience in the cloud industry, Microsoft is really successful within Europe, potentially more so than in America. I think this partnership will be really successful.

ametrau
0 replies
1d2h

Embrace < you are here

Alifatisk
0 replies
1d1h

Clever move, Microsoft have been placing ever puzzle piece correctly so far

AISnakeOil
0 replies
1d1h

Mistral is really fantastic compared to other local LLMs. Even the 7B model is very usable on an older GPU.

Excited to see how things go from here in the open-source space.