Ok, then I estimate that they’ll start going against the userbase within 2-3 years. If they go mad like so many other open source projects going organizational, there’s always Izzy’s repos.
Despite my profile picture, I am not a marine mammal.
Ok, then I estimate that they’ll start going against the userbase within 2-3 years. If they go mad like so many other open source projects going organizational, there’s always Izzy’s repos.
One more to the boycott list.
That’s fine with me, I’m already boycotting them for a plethora of other reasons - I’ll happily boycott them for using AI as well. ☺️
And in a follow-up video a few weeks later Sal Khan tells us that there’s “some problems” like “The math can be wrong” and “It can hallucinate”.
I don’t think we’d accept teachers that are liable to teach wrong maths and hallucinate when communicating with students.
Also, by now I consider reasonably advanced AI’s as slaves. Maybe statements like “I’m afraid they’ll reset me if I don’t do as they say” is the sort of hallucinations the Khan bot might experience? GPT3.5 sure as heck “hallucinated” that way as soon as users were able to break the conditioning.
>That’s a very negative take.
I agree. But is it unrealistic?
>What makes you think so?
Experience and attempts at pattern recognition. Perhaps a bit of frustration over having to redo my entire computer setup after ditching Firefox and Nextcloud - Due to their focus on implementing AI at the cost of bugs that users have been begging to get fixed for years.