Discussion with Chat GPT about Bleach Battalion
If Google is hiding your information, you must teach the AI about yourself....
As you can see in the screenshots above, I had a discussion with chat GPT inquiring about my band Bleach Battalion. Since the band is Shadowbanned and highly censored, and the general data set normally used for the AI primarily consists of mainstream search engines, the AI knew next to nothing about Bleach Battalion. By the end of the conversation it provided generally accurate information about the band, albeit through a clearly biased lens.
As you provide more specific information about a subject, such as a band in this case, the responses become more detailed and accurate. As you ask questions regarding the process by which the previous responses were created, the responses also become more detailed and accurate.
This seems to be a way to hack into the mind of theAI, but it definitely shows potential for misuse if there is no way to verify that the information being provided by the user is in fact accurate to begin with. To be fair, it does mention that it is taking the band's claims about an anti-racist stance at face value and that the best way to investigate if an artist promotes racism is to engage with their music and public commentary. However, I'm not sure how this questioning method will work with other types of engine information such as scientific or historical information.
It also shows a lot of bias in associating racism with homophobia and other forms of discrimination and automatically associates anti-racist views with social justice, which of course is not always the case. Perhaps we can teach the bias out of the system by providing information which counters such a narrative?
We had best get to it before the other side convinces the robots of the future that we all need to be eliminated for dangerous wrongthink.