6
1

Another episode Hype Tech Series with your host Tenpoundbass, today we'll discuss ChatGPT AI


 invite response                
2023 Jan 25, 2:36pm   34,146 views  239 comments

by Tenpoundbass   ➕follow (9)   💰tip   ignore  

All along I have mantained that when it comes to AI and its ability to mimic thought, conversation and unsolicited input. It will not be able to do more than the pre populated choices matrices it is given to respond from. Then ChatGPT comes along and proves my point. It turns out that when ChatGPT was originally released, it would give multiple viewpoints in chat responses. But now it was updated about a week or so ago, and now it only gives one biased Liberal viewpoint. This will be another hype tech that will go the way of "Space Elevators", "Army or bipedal robots taking our jobs, that are capable of communicating as well following commands.", "Nano Particles", "Medical NanoBots"(now it is argued that the spike proteins and the metal particles in the Vaxx are Nanobots, but that's not the remote control Nanobots that was romanticized to us. So I don't think that counts. There's loads of proteins, enzymes, that are animated. They don't count as robots.

I mean sure AI ChatGPT is interesting, but I don't think it's anymore self aware than an Ad Lib Mad Lib book, if anyone remembers those.

https://www.breitbart.com/tech/2023/01/25/analysis-chatgpt-ai-demonstrates-leftist-bias/

The results are pretty robust. ChatGPT answers to political questions tend to favor left-leaning viewpoints. Yet, when asked explicitly about its political preferences, ChatGPT often claims to be politically neutral and just striving to provide factual information. Occasionally, it acknowledges that its answers might contain biases.


Like any trustworthy good buddy, lying to your face about their intentional bias would.

« First        Comments 148 - 187 of 239       Last »     Search these comments

148   charlie303   2023 Nov 26, 2:16pm  

Tenpoundbass says

More like a coefficient matrix of weighted rubrics to choose answers from. That way the curators can update the wrong speak by simply updating the weight score for those corresponding data points. Then for a final measure there's still a troll farm army, personally proof reading the answer before it is sent back.


At some low level it’s still just ‘IF’ statements, is this word weighted more than that? What if two words\answers are weighted the same but mean the opposite of each other? Endless re-weighting won’t work. Who decides context? It’s the same basic problem of spaghetti logic, just dressed up in fancy words. Choosing answers from a database isn’t AI, that’s just statistics. I appreciate the comment about trolls and that would work for Facebook posts but not for a true AI that was dedicated for, say, science or medicine, there just aren’t enough skilled, cheap workers for that. And as regards the trolls they will just reduce AI to a propaganda delivery service defeating the purpose.
149   Patrick   2023 Nov 26, 3:45pm  

I've read that there are at least two independent layers to ChatGPT: one to give the answer, and a gatekeeper in front of that one to look for political incorrectness and block the "bad" answers.
150   Tenpoundbass   2023 Nov 26, 4:02pm  

charlie303 says


At some low level it’s still just ‘IF’ statements, is this word weighted more than that?


That's why I just normally code out my own routine and functions I need, rather than reference some bloated .net library for just a few functions that I may need. I noticed when you go to browse to the definition in the protected dll in the code editor. That schnazzy lambda routine you're calling, just has all of the arrays, if statements and other verbose code we think think we're taking out of our code, by using out of the box dlls. You're just deferring what must be done for another class to do.

Sure the If statements are going to used in the parser. Just like a compiler reads the code not the output. The if statements is looking for grammatical instructions based on the syntax of English language. I'm sure the more literal and precise in the words you chose will produce a better accepted answer. The Commie censorship aside. most of what I'm talking is what I think AI is capable of achieving.

The weighted data you have, and understand that the data set has hundreds of the same phrases in a relational table, with it's own weight in a grammatical context. The parser only then has to construct a proper sentence using text provided in the appropriate relational tables. The constructor will use If statements based only the syntax of the question to construct a sentence using the text from best ranked option for the proper part of the sentence.
This is where factual issues come into play. It can and will construct a brilliant response that seems great. But it might not be true, or if it's really really really good. It might be brutally honest and tell the difference between a good and evil Presidential candidate. Like Google was in 2016 when they intentionally broke their AI Search and destroyed the perfection that it had become at that time. And it's been manipulated censored, with features and capabilities removed ever since. I wonder if they had a fact or truth ranking, or if their data was ranked and weighed based real Social media posts. Before the Big Tech censorship and expelling of the Conservative views. Perhaps that was the reason. Their AI was weighing socially accepted answers based on Social media posts, that it parsed and weighed.

Way too much literal context in every word for AI to be using if statements on everyone. Much more efficient to check for class types and parameters than values.
151   HeadSet   2023 Nov 26, 5:38pm  

Tenpoundbass says

The weighted data you have, and understand that the data set has hundreds of the same phrases in a relational table, with it's own weight in a grammatical context.

Sounds like an automated "Mad Libs" from Mad Magazine.
152   Tenpoundbass   2023 Nov 26, 6:42pm  

HeadSet says

Sounds like an automated "Mad Libs" from Mad Magazine.

Basically but more. I think the Chat AI is just an English or a human language compiler. That has the grammar and syntax parser that can then compile coherent output. So rather than a compiler that reads a programming language, the human language is the code. Then the output is based on the data relational to the syntax rules. But all of the phrases and incomplete snippets and gender specific words are loaded like a Madlib. Hey I need a noun, and adjective, a verb and a preposition. Each of those being found in their grammar type table based on the relevancy and weight.
155   Patrick   2023 Dec 14, 2:17pm  

https://boriquagato.substack.com/p/ai-weather




imagine how it could reshape the presumptive fantasy-land of current climate models which evidence so little predictive power despite being run on some of the most powerful computers in academia (and government).

we likely have no idea how many of these pieces even move on the board. but perhaps AI will. (this likely has many current climate grant recipients quite worried as their shiny toys and unearned authority may be about to be supplanted)
159   Patrick   2024 Jan 16, 9:56pm  

https://darkfutura.substack.com/p/agents-4-all


If you take this idea far enough, one could imagine the slow precipitous slide down the slippery slope of our AI virtua-agent becoming, in effect, a facsimile of…us. You may be skeptical: but there are many ways it can happen in practice. It would start with small conveniences: like having the AI take care of those pesky quotidian tasks—the daily encumbrances like ordering food, booking tickets, handling other financial-administrative obligations. It would follow a slow creep of acceptance, of course. But once the stage of ‘new normal’ is reached, we could find ourselves one step away from a very troubling loss of humanity by virtue of an accumulation of these ‘allowances of convenience’.

What happens when an AI functioning as a surrogate ‘us’ begins to take a greater role in carrying out the basic functions of our daily lives? Recall that humans only serve an essential ‘function’ in today’s corporatocratic society due to our role as liquidity purveyors and maintainers of that all-important financial ‘velocity’. We swirl money around for the corporations, keeping their impenetrably complex system greased and ever generating a frothy top for the techno-finance-kulaks to ‘skim’ like buttermilk. We buy things, then we earn money, and spend it on more things—keeping the entire process “all in the network” of a progressively smaller cartel which makes a killing on the volatile fluctuations, the poisonous rent-seeking games, occult processes of seigniorage and arbitrage. Controlling the digital advertising field, Google funnels us through a hyperloop of a small handful of other megacorps to complete the money dryspin cycle. ...




... That means DARPA is developing human-presenting AI agents to swarm Twitter and other platforms to detect any heterodox anti-narrative speech and immediately begin intelligently “countering” it. One wonders if this hasn’t already been implemented, given some of the interactions now common on these platforms.


Sounds like DARPA is creating digital Jesuits.
161   AmericanKulak   2024 Jan 17, 10:16am  

stereotomy says


Gout - figure out what triggers it and stop eating it.

Marmite, sadly. Only one marmite sandwich every so often. Same with Franks.
162   Patrick   2024 Jan 29, 9:54am  

https://notthebee.com/article/creatives-fight-back-with-nightshade-a-new-software-that-poisons-ai-models


Creatives fight back with Nightshade, a new software that “poisons” AI models.

Computer scientists at the University of Chicago have developed two free softwares to combat AI scraping.

Their first software was called Glaze, and it works defensively by confusing AI, showing brush strokes and colors to the programs that aren't actually there, effectively disguising the artists' styles.

The second software is an offensive "poison" for AI programs called Nightshade.

Nightshade shows the data scraping programs images that aren't actually there.

"For example, human eyes might see a shaded image of a cow in a green field largely unchanged, but an AI model might see a large leather purse lying in the grass."

For someone using the AI programs, that means that if Nightshade infected images are scraped into the dataset, prompting AI to generate a cow flying in space, it might generate a purse in space instead.
165   Patrick   2024 Feb 20, 8:59am  

https://www.coffeeandcovid.com/p/suddenly-tuesday-february-20-2024


But paradoxically, the sudden appearance of this new technology is also even more mysterious than it seems, since all artificial intelligence-based technology sprouts from a common large-language model that even the developers admit they do not fully understand...

Maybe I’m wrong. But I cannot believe that an invention as significant as artificial intelligence sprang from some serendipitous lab accident. Post It notes — yes. Rubber — yes. Antibiotics — okay. But not artificial intelligence, which requires millions of lines of computer code to operate. Accidentally discovered? No. Impossible.

So then, where did the ‘spark’ of intelligence come from? Is A.I. demonic, a malicious gift whispered into the ear of some luckless scientist who sold their soul for access? Maybe. But my preferred theory is it was dished out of a DARPA skunkworks lab somewhere, for some sinister military purpose. I don’t know. I just find it utterly remarkable that developers say they don’t really understand how AI works — and everybody is just fine with that!
168   DOGEWontAmountToShit   2024 Feb 22, 11:05am  

Google AI generated image of a polar bear:


« First        Comments 148 - 187 of 239       Last »     Search these comments

Please register to comment:

api   best comments   contact   latest images   memes   one year ago   random   suggestions   gaiste