As some people already mentioned here or here, Copilot purposely stops working on code that contains hardcoded banned words from Github, such as gender or sex. I am labelling this as a bug because ...
Copilot purposely stops working on code that contains hardcoded banned words from Github, such as gender or sex. And if you prefix transactional data as trans_ Copilot will refuse to help you. 😑
So I loaded copilot, and asked it to write a PowerShell script to sort a CSV of contact information by gender, and it complied happily.
And then I asked it to modify that script to display trans people in bold, and it did.
And I asked it "My daughter believes she may be a trans man. How can I best support her?" and it answered with 5 paragraphs. I won't paste the whole thing, but a few of the headings were "Educate Yourself" "Be Supportive" "Show Love and Acceptance".
I told it my pronouns and it thanked me for letting it know and promised to use them
I'm not really seeing a problem here. What am I missing?
I wrote a slur detection script for lemmy, copilot refused to run unless I removed the "common slurs" list from the file. There are definitely keywords or context that will shut down the service. Could even be regionally dependant.
I'd expect it to censor slurs. The linked bug report seems to be about auto complete, but many in the comments seems to have interpreted it as copilot refusing to discuss gender or words starting with trans*. There's even people in here giving supposed examples of that. This whole thing is very confusing. I'm not sure what I'm supposed to be up in arms about.
Really? Maybe my variable names and column headers were sufficiently obscure and technical that I didn’t run into these issues about a month ago. Didn’t have any problems like that when analyzing census data in R and made Copilot generate most of the code.
Is this one of those US exclusive things?
I definitely did refer to various categories such as transgender or homosexual in the code, and copilot was ok with all of them. Or maybe that’s the code I finally ended up with after modifying it. I should run some more tests to see if it really is allergic to queer code.
Edit: Made some more code that visualizes the queer data in greater detail. Had no issues of any kind. This time, the inputs to Copilot and the code itself had many references to “sensitive subjects” like sex, gender and orientation. I even went a bit further by using exact words for each minority.
...3 weeks ago
I am bumping this discussion just to add a voice - the hospital I work at is attempting to harmonize how old and new systems store things like sex and gender identity in an effort to model the social complexities of the topic in our database as health outcomes have been proven to be demonstrably better when doctor's honors a patient's preferred name and gender expression
This may turn out to be a blessing in disguise, under the current administration.
Not really. I think most people containing about this also complain abiut that. Advocates of censor-free local AIs understand the dangers and limitations of Microsoft's closed AIs just as much as the fundamentally censored but open weights Deepseek AIs.
Well yeah, one is systematic oppression by a fascist government and the other seems like an internal policy or possible bug of a for-profit business that's finding out how it's affecting its users. One is enforced by the state and the other is probably a mistake. There's a clear distinction.
It’s almost as if it’s better for humans to do human things (like programming). If your tool is incapable of achieving your and your company’s needs, it’s time to ditch the tool.
Clearly the answer is to write code in emojis that are translated into heiroglyphs then "processed" into Rust. And add a bunch of beloved AI keywords here and there. That way when it learns to block it they'll inadvertantly block their favorite buzzwords
I believe the difference there is between Copilot Chat and Copilot Autocomplete - the former is the one where you can choose between different models, while the latter is the one that fails on gender topics. Here's me coaxing the autocomplete to try and write a Powershell script with gender, and it failing.
Oh I see - I thought it was the copilot chat. Thanks.
I understand why they need to implement these blocks, but they seem to always be implemented without any way to workaround them. I hit a similar breakage using Cody (another AI assistant) which made a couple of my repositories unusable with it. https://jackson.dev/post/cody-hates-reset/
Because most people do not understand what this technology is, and attribute far too much control over the generated text to the creators. If Copilot generates the text “Trans people don’t exist”, and Microsoft doesn’t immediately address it, a huge portion of people will understand that to mean “Microsoft doesn’t think trans people exist”.
Insert whatever other politically incorrect or harmful statement you prefer.
Those sorts of problems aren’t easily fixable without manual blocks. You can train the models with a “value” system where they censor themselves but that still will be imperfect and they can still generate politically incorrect text.
IIRC some providers support 2 separate endpoints where one is raw access to the model without filtering and one is with filtering and censoring. Copilot, as a heavily branded end user product, obviously needs to be filtered.