743: The Current AI Moment

3 Likes

Great episode!

Three things:

If you are averse to paying for any AI services, brush up on your web searching skills and you can find beta or free versions of just about any AI service right now.

May not stay free and may have more usage limits added, but in these early days, all these companies want exposure and eyeballs and are not (yet) focused on monetization and long-term profitability.

Related - The hosts appear to have not tried Microsoft Copilot that is part of Bing. I started using it almost a year ago because of three benefits:

It is totally free (unlike paying OpenAI for an account)
It uses the latest version of ChatGPT from OpenAI
Microsoft annotates every result with footnotes that link to the original source material.

Maybe not as polished as the service mentioned in the podcast, but having both an AI generated search result as a narrative that includes all the source links has been working great for me.

Also, it may not be obvious, but you can use it as a general purpose generative AI, not just for search.

Any of the prompts you can submit to ChatGPT you can type into CoPilot - not just search requests.

9 Likes

Thank for yet another enjoyable show!

Some valuable insights on the “why can’t ai be an all out personal assistant yet” by Cal Newport can be found as podcast here or as written word here

Furthermore, if you’re a Setapp subscriber: check out superwhisper as alternative to MacWhisper

Fun episode. I really want to hear about Sparky’s use of an LLM to summarize Readwise highlights.

On the Copilot Coding front - it’s far from rosy. I’m already helping clients sort out this mess. Linked is a research paper that documents how Copilot is helping to make code quality and maintainability worse. Coding on Copilot: 2023 Data Suggests Downward Pressure on Code Quality (incl 2024 projections) - GitClear

1 Like

Yep, the “easy code” that has low quality and is replicated all over the place is not really the code I want generated.
Instead I focus on Copilot Chat for help with “what does this function do” after I’ve written it, tests (of course) and documentation.
Of course, for “line completion” it also does a great job, but I mostly keep it at the line level, and don’t let it create entire functions.

1 Like

Could we please clone you and spread this knowledge around. I see many devs, who feel pressured to produce more. They use Copilot to help them write more ‘code’.

2 Likes

Agreed. I downloaded the Microsoft Edge browser and use it for the sole purpose of running Copilot within the browser. It’s pretty good, and I like that it gives references.

2 Likes

I tried perplexity (free-tier) straight after listening. It’s very “user-friendly” and well-presented and often gave some surprisingly useful answers as the hosts said.

For example, I asked it to compare two specific smart telescopes - something I have been researching (the old-fashioned way) recently. I was very impressed with the first summary it gave. It picked up the key points perfectly, succinctly and accurately, with the footnote style references, which cited a lot of the better quality and more in-depth sites I had found. I genuinely think it did a better job than I would have done, and it produced it instantly.

I then asked for some more specifics (e.g. pros and cons of each) and began to notice out-of-date and even just plain wrong material creeping in. The errors weren’t just coming from out of date material being cited, but confusing e.g. specifics of particular models with previous models in the same range. Perplexity is using existing LLMs in the background, so this is not a surprise. As the hosts said, “hallucinations” happen quite a lot.

But perplexity presents the material so well, and so accurately in context, that it is jarring to realise that it is being “a perfect con artist”: hiding significant errors and misinformation amid impressive true results and presenting everything with persuasive confidence (and even humility) that is likely to give it credibility.

I liked the analogy of the intern, but although a human intern might not even know what they don’t know yet, so would need their work and understanding checked, they would be unlikely to be so impressive and persuasive while being wrong - and if they are, they would quickly be seen as dangerous and a mentor would try to get them to change.

I came to a different conclusion than David and Stephen: I thought perplexity was LESS good than many other AI front-ends. I like it giving references and citations, and it “writes” well, but it’s so credible that it really does have to be better at not presenting false information, or at least indicating how MUCH good evidence (or not) there is behind each point. The number and quality of citations making the same point matters at least as much as there being a citation.

3 Likes

Do you check the references? Beware sometimes the facts it gets from the references are made up.

2 Likes

Does anyone have the link to the official OpenAi app for Mac?

I can’t seem to find it on the App Store or otherwise

Looks like it is not widely released yet:

https://www.reddit.com/r/OpenAI/comments/1cr5g5z/desktop_macos_app/

Someone commented 15 days ago (as of this post):

You can download the app, but as u/princesspbubs pointed out, your account has to be activated server-side, or you receive this:

1 Like

I got access last week.

An offer to download it suddenly popped up at the top of the page while using ChatGPT in safari on my Mac.

I now have an option to download the app from the settings, while using chat got on safari in the Mac.

It’s a nice little app, with plenty of room to improve.

Were you signed up for it somehow?

Oh, yes, good question: you have to be a paying customer ($20 a month) to use the Mac app.

1 Like

I just listened to this episode and I was a little surprised that there was not even any mention of Grok. I understand that the hosts have issues with X, but they have issues with Meta, and they still mentioned their AI. Regardless of your feelings about X, I believe Grok will be a significant presence in AI.

The brilliant part was—as Sparky suggested—Apple branding this their own way.

“Apple Intelligence” (brilliant)

Now, if an Apple spokesperson or a marketing rep accidentally refers to “A.I.” and is called out on that they can reply:

“It stands for ‘Apple Intellegence’. You knew that, right? Like, doesn’t everybody?”

I find the marketing behind Apple’s AI to be brilliantly interesting.

They went from missing the boat to leading the pack with a few clever moves:

Branding AI as “Apple Intelligence”. Corny, but it works to position them as the leader/trendsetter.

Instead of competing directly with ChatGPT, Grok, CoPilot, Meta, and others slinging it out to be the LLM winner, they lapped all of them by introducing their own on-device AI that merges Apple’s passion for privacy and security with the ability to devour personal data and provide more intelligent assistance.

Imagine if Google or ChatGPT try to offer a tool that has to suck up your calendar, contacts, email, and notes to achive the same level of personalization?

The coup de grâce was encapsulating ChatGPT in a plug-in “terminal window” complete with anti-tracking session encryption and annonymization and then proclaiming that all other LLM’s could also be “just another plug-in”.

They Sherlocked all the leading AI’s by positioning them as tools feeding Apple AI instead of peers or direct competitors.

As MacSparky said on the last podcast, “Now, they just have to deliver!”

5 Likes

Does announcing features, some of which are not expected until sometime in 2025, move you to the head of the pack? And even then will not be available to more than 90% of current iPhone users unless they have purchased a new phone and speak US English.

Apple may indeed end up winning the game, but currently has yet to sit down at the table.

2 Likes

Okay, I’ll bite. Why?

For the same reason that Meta’s or Google’s AI will be. It is a huge platform with millions of active users, so it will continue to build an enormous knowledge base. I don’t know that any of them will be able to catch OpenAI, just because of the head start they have, and now their relationships with the two biggest operating systems, but I still think all three (Meta, Google, Grok) will be significant.

1 Like