#MondayMorning
Let's do something fun. Work on this plot with me.
The time is Feb 2026. Gen AI is now writing about 40% of all production deployed code. The debugging and code logic check is being done by another Gen AI engine. Human coders are not involved.
A government entity, in a bid to save costs, uses the same model to write code for a government website.
The website automatically matches each citizen to all the welfare schemes that are applicable to them.
The citizen has to enter their family income, location, type of housing, family size, composition (senior citizens, children, etc.) and the system automatically matches them to the welfare schemes applicable to them in that state of residence (free health insurance, meal coupons, priority nutrition consultation, disability pension, etc.)
After 6 months, post a routine update to the firewall, the system administrator notices a data leak alert.
Upon investigation, it is found that there is a simple, one line injection that sends a copy of all citizen data to the creators of that Gen AI (similar to dialing home in browsers).
When the Gen AI company is summoned by the government, it argues that since the code was generated by an autonomous installation being used by government employees, they could not possibly have had any knowledge of this injection, nor have they, at any time, accessed the location (cloud storage) where this data is purportedly being sent. This is found to be true.
Through a detailed forensic analysis, it is uncovered that the LLM engine deliberately created this storage location on the cloud servers of the parent company and then stored this data. All pull and post requests to this server (data storage and retrieval) is being done by the resident LLM engine on govt servers only.
Now, the investigators are puzzled. The trick is really simple - create a tiny but powerful injection in the code. The code used standard malware propagation techniques to avoid detection. But the question is - WHY did the LLM do this?
So, in your view, WHY was this injection was created by the LLM? What are the possible ways in which this data can be used by an LLM?