Submitted by Fit-Meet1359 t3_110vwbz in singularity
amplex1337 t1_j8dl4d3 wrote
Reply to comment by chrisc82 in Bing Chat blew ChatGPT out of the water on my bespoke "theory of mind" puzzle by Fit-Meet1359
It doesn't understand anything, it's a chatbot that is good with language skills which are symbolic. Please consider it's literally just a GPU matrix that is number-crunching language parameters, not a learning, thinking machine that can move outside of the realm of known science that is required for a doctorate. Man is still doing the learning and curating it's knowledge base. Chatbots have been really good before chatGPT as well.. you just weren't exposed to them it sounds like
OutOfBananaException t1_j8hiqoo wrote
Given me one example of an earlier chatbot that could code in multiple languages.
amplex1337 t1_j8qp89h wrote
chatGPT doesn't understand a thing it tells you right now, nor can it 'code in multiple languages'. It can however fake it very well. Give me an example of truly novel code that chatGPT wrote that is not some preprogrammed examples strung together in what seems like a unique way to you. I've tried quite a bit recently to test its limits with simple yet novel requests, and it stubs its toe or falls over nearly every time, basically returning a template, failing to answer the question correctly, or just dying in the middle of the response when given a detailed prompt, etc. It doesn't know 'how to code' other than basically slapping together code snippets from its training data, just like I can do by searching in google and copy pasting code from the top results from SO etc. There are still wrong answers at times.. proving it really doesn't know anything. Just because there appears to be some randomness to the answers it gives doesn't necessarily make it 'intelligence'. The LLM is not AGI that would be needed to actually learn and know how to program. It uses supervised learning (human curated), then reward based learning (also curated), then a self-generated PPO model (still based on human-trained reward models) to help reinforce the reward system with succinct policies. Its a very fancy chatbot, and fools a lot of people very well! We will have AGI eventually, its true, but this is not it yet and while it may seem pedantic because this is so exciting to many, there IS a difference.
OutOfBananaException t1_j8qu042 wrote
I never said it 'knows' or displays true intelligence, only that it performs at a level far above earlier chatbots that didn't come close to this capability.
Representative_Pop_8 t1_j8iqgft wrote
what is your definiton of understand?
what is inside internally matters little if the results are that it understands something. The example shown by OP and many more, including my own experience clearly shows understanding of many concepts and some capacity to quickly learn from interaction with users ( without needing to reconfigure nor retain the model) though still not as smart as an educated humans.
It seems to be a common misconception , even by people that work in machine learning to say these things don't know , or can't learn or are not intelligent, based on the fact they know the low level internals and just see the perceptions or matrix or whatever and say this is just variables with data, they are seeing the tree and missing the forest. Not knowing how that matrix or whatever manages to understand things or learn new things with the right input doent mean it doesn't happen. In fact the actual experts , the makers of these AI bots know these things understand and can learn, but also don't know why , but are actively researching.
>Man is still doing the learning and curating it's knowledge base.
didn't you learn to talk by seeing your parents? didn't you go years to school? needing someone to teach you doesn't mean you don't know what you learned.
RoyalSpecialist1777 t1_j8kwg8b wrote
I am curious how a deep learning system, while learning to perform prediction and classifation is any different than our own brains. It seems increasingly evident that while the goals used to guide training are different but the mechanisms of learning effectively the same. Of course there are differences in mechanism and complexity but what this last year is teaching us is the artificial deep learning systems work to do the same type of modeling we undergo when learning. Messy at first but definitely capable of learning and sophistication down the line. Linguists argue for genetically wired language rules but really this isn't needed - the system will figure out what it needs and create them like the good blank slates they are.
There are a lot of ChatGPT misconceptions going around. For example that it just blindly memorizes patterns. It is a deep learning system (very deep) that, if it helps with classification and prediction, ends up creating rather complex and functional models of how things work. These actually perform computation of a pretty sophisticated nature (any function can be modeled by a neural network). And this does include creativity and reasoning as the inputs flow into and through the system. Creativity as a phenomena might need a fitness function which scores creative solutions higher (be nice to model that one so the AI can score itself) and of course will take awhile to get down but not outside the capabilities of these types of systems.
Anyways, just wanted to chime in as this has been on my mind. I am still on the fence whether I believe any of this. The last point is that people criticize ChatGPT for giving incorrect answers but it is human nature to 'approximate' knowledge and thus incredibly messy. Partially why it takes so long.
Caring_Cactus t1_j8izqb9 wrote
Does it have to be in the same way humans see things? It's not conscious, but it can understand and recognize patterns, is that not what humans early on do? Now imagine what will happen when it does become conscious, it will have a much deeper understanding to conceptualize new interplays we probably can't imagine right now.
Viewing a single comment thread. View all comments