This is not investment advice. The author has no position in any of the stocks mentioned. Wccftech.com has a disclosure and ethics policy.
AI singularity is defined as an epochal threshold where artificial intelligence (AI) demonstrably exceeds human intelligence, unlocking productivity nirvana in the process. Now, with the release of OpenAI’s GPT-4 Turbo, humanity has come a significant step closer to realizing this ultimate utopic dream.
OpenAI unveiled its most powerful AI-based Large Language Model (LLM) yesterday, creating a sizable buzz across the global tech sphere. Dubbed the GPT-4 Turbo (Generative Pre-Trained Transformer-4 Turbo), the production-ready version is set to officially debut in the next few weeks. While the vanilla GPT-4 was trained on information dating up to September 2021, the GPT-4 Turbo is equipped with much more recent data, having been trained on information dating up to April 2023.
As far as the salient features of GPT-4 Turbo are concerned, OpenAI noted in its first-ever developer conference yesterday:
- Two versions, with one specializing in analyzing text, while the other is capable of processing and interpreting textual as well as image-based context. The former is now available via an API Preview.
- The new AI model would be considerably cheaper for developers to operate, with input costing only $0.01 per 1,000 tokens (a basic unit of text or code that an LLM uses to generate its response) vs. $0.03 in the case of the vanilla GPT-4 model. In either case, output would cost around $0.03 per 1,000 tokens.
- OpenAI’s latest offering will continue to support text-to-speech requests, image-based input, and image generation requests via DALL-E 3 integration.
- GPT-4 Turbo is reportedly capable of handling highly complicated tasks encased within a single prompt, courtesy of its much more exhaustive training.
- The single biggest change, however, relates to the LLM’s token context window, a measure of the richness or depth of input and output of a particular LLM. GPT-4 Turbo offers a context window that is around 4x that of GPT-4, capable of processing 128,000 tokens. For context, 1,000 tokens are roughly equivalent to 750 words. This means that OpenAI’s latest offering is capable of processing 96,000 words. Anthropic’s Claude 2 can only process 100,000 tokens right now.
Yesterday OpenAI released GPT-4
This is the largest step towards Singularity we’ve seen in our Lifetime
Here’s why
— Joe Ayoub (@Crypto_Joe10) November 7, 2023
This brings us to the crux of the matter. The co-founder of rug.ai and Diffusion, Joseph Ayoub, believes that the release of GPT-4 Turbo is “the largest step towards Singularity we’ve seen in our lifetime.”
Context is one of the largest constraints of intelligent computation.
Without larger context, models are unable to keep track of broader data-sets, with more complex problems.
This improvement means now you can input huge datasets and get more intelligent answers…
— Joe Ayoub (@Crypto_Joe10) November 7, 2023
Ayoub rightly identifies context as one of the primary constraints vis-a-vis artificial intelligence. One of the fundamental reasons why humans are much smarter than machines is, of course, our ability to grasp and understand significantly greater contextual information.
But with increased context increases a model’s cognitive ability increases by multiples with Long-Term memory.
This upgrade facilitates models that create THEIR OWN programs to solve tasks
All that means is that models can more efficiently solve HARDER problems by themselves
— Joe Ayoub (@Crypto_Joe10) November 7, 2023
What’s more, GPT-4 Turbo allows users to create custom GPTs! This means that you can optimize the LLM based on your unique needs and requirements.
Do you think we are on the precipice of an AI singularity? Let us know your thoughts in the comments section below.