Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

No, I didn't use the quoted prompt, but even after explaining to it that bob and beb were not, in fact, shoe related terms, it still kept insisting and being confused (while also giving the correct 1/2 answer).

It can do it, but its not deterministic, and it doesnt really do it well. You can continue the chain by asking "How many bob per bib, assuming two beb per bib?", and see if it chokes then. It sometimes does, sometimes doesnt.



GPT-4:

   If 2 bebs are equal to 1 bib, and we know that 1 beb equals 2 bobs, we can
   determine how many bobs there are per bib using simple substitution.
   
   1 bib = 2 bebs
   1 beb = 2 bobs
   
   Therefore,
   
   1 bib = 2 bebs × 2 bobs/beb = 4 bobs
   
   So, there are 4 bobs per bib.
Nitpick: A properly done substitution would've arrived at

   1 bib = 2 × (2 bobs)
without needing any of the "2 bebs × 2 bobs/beb" nonsense. It doesn't teach this task very well.


You do realize that the current implementations get their context polluted by your prior conversation, right?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: