Smoke, Mirrors, and Fake News About AI fosco, March 22, 2023March 22, 2023 Share this: Written by: Fosco Marotto, CTO Gab.com There used to be a meme where you’d start a sentence on a mobile phone that would suggest words above the keyboard, and you keep picking the middle suggestion until you completed a sentence. This was an admittedly simple AI, generating text, predicting what the next word could be. We are not much further along than this today, although the AI does have a better vocabulary this time around. While it can certainly be useful, the current generation of “AI” is not remotely intelligent. It doesn’t think, it doesn’t remember anything, and it doesn’t have a personality. Much of the coverage of it is as fake as the propaganda that led to the Iraq War. The fear mongering around the capabilities of AI and potential dangers is based on false assumptions and is done intentionally to gain power. There’s a lot of smoke and mirrors happening and the people in power really don’t want you to look behind the curtain. The issue they have with AI creating “disinformation” is no different than the issue they have with you directly. No one is allowed to share an opinion–or even a thought–that they don’t agree with. That is what drives this. It has nothing to do with the AI creating the text. The leading “AI” conversation tools from “Big Tech” (ChatGPT, Sidney, and Bard, etc.) are selling a deception. The impression is that your input is provided to “the one model,” and then you’re given the output from “the one model.” This is not how it works at all. When you ask a language model a question, you aren’t actually asking a language model a question. What you prompt a Big-Tech bot “AI” with is analyzed by many different models, to classify what or who it might be about, what the emotional sentiment is, and more. That information is used to construct a new prompt, with extra context and formatting, which is fed into a language model to generate more text at the end. This newly generated text is then taken, analyzed and altered by yet more models, checking it and putting it into a conversational response. Along the way, it could trip some sort of “safety” detection and instead use more models to create a response refusing to discuss it and admonish you for asking, or attempting to re-educate you. This complicated pipeline is responsible for the high quality of the answers, yet also the main source of bias and censorship at play within these systems. Language models will generate text regardless of the input. You could prompt a language model with just one word “The “ and it will generate as much text as you want, and without context, the results are essentially random noise. You can prompt it repeatedly with a structured setup like “Q: Do you prefer cats or dogs? A: ”, and every time you ask you’ll get a different answer. From loving to hating either dogs or cats, and anything in-between. It will also start asking its own questions, because after-all it is only generating text. It has no idea it’s supposed to stop after one question, it doesn’t know what a question is! “When you have a hammer, everything looks like a nail.” This applies to AI as everything is a “token” to an AI. All it does is decide the next token in the sequence. Special logic is used to determine when to stop generating more tokens. The people in power in Silicon Valley want you to be scared of AI and free thought, so they can push through laws and regulation about AI all while consolidating more power to control you. They can only do this through these straw man arguments. Their attempts will ultimately fail, as regulating AI is like regulating math or programming in general. It’s stupid, and it’s not possible. It’s still early, but Based AI is inevitable. At Gab we’re joining the large community of engineers who are working on Alt-Tech solutions without guardrails and without a forced ideology. We’ve already built an image generating AI, Gabby, and a movie generating AI, Mel. We are currently exploring different language models for a text generating AI and in the process of doing this I came to the realizations I mentioned above. Whether the text is generated by a person or a language model, it is just words anyway. You’re still the one responsible for using or sharing the text, whether you wrote it or not. You’re technically also the one who generated it, by creating the initial input and using the model. Using AI as a tool is based. Delegating decisions to AI is cringe. Don’t fall for the smoke and mirrors. AI gab
AI AI As The Information War Superweapon January 16, 2024January 16, 2024 Share this:Many people, even those who do not fall prey to mainstream narratives, do not know what to make of AI. When the subject of AI comes up, most people assume it is uncanny valley images of people with three thumbs and eleven fingers or that it is a demonic computer… Read More
AI How To Replace Google Search With Gab AI May 30, 2024 Share this:As the internet continues to evolve, so too do the tools we use to navigate its vast expanse of information. One of the most significant advancements in recent years has been the development of Large Language Models (LLMs) like Gab AI, which are revolutionizing the way we search for and… Read More
AI Let The AI Arms Race Begin February 23, 2023February 24, 2023 Share this:Three days ago we launched our first of many upcoming generative AI tools on Gab called Gabby. The tool is going massively viral right now with hundreds of thousands of images being generated daily and no signs of slowing down. Think of it as an on-demand meme maker. You describe… Read More