
Photo by Cookie the Pom on Unsplash
The next generation of AI Assistants may seek extensive access to our personal data, accounts and devices features. Can we trust their developers to protect our privacy and security?
Photo by Cookie the Pom on Unsplash
“Hey [enter AI assistant name here], can you book me a table at the nearest good tapas restaurant next week, and invite everyone from the book club?” Billions of dollars are invested in companies to deliver on this. While this is a dream that their marketing departments want to sell, this is a potential nightmare in the making.
Major tech companies have all announced flavours of such assistants: Amazon’s Alexa+, Google’s Gemini inspired by Project Astra, Microsoft’s Copilot AI companion and Apple’s Siri powered by Apple Intelligence, and all of them aiming to become a more useful assistant by integrating deeply with your apps, accounts and real life.
AI Assistants need to access apps, data and device services in order to deliver on their promise to operate as agents capable of doing work for us. So in booking that table at a restaurant in the example above, or using your location and your calendar to inform people that you are running late, AI assistants need access to apps (e.g: your messaging or email app), data (e.g.: calendar details and contacts), and device services (e.g. your current location).
This is a significant change from voice assistants today. At the moment they help you interact with specific apps and services (“play me this song”, “what is the weather today?”. From a security and privacy perspective, they are usually quite constrained.
Integrating distinctly managed services into one requires more than what voice assistants of recent years have been able to provide. In the words of Amazon’s SVP of Devices and Services about Alexa+: “It wasn’t just rearchitecting [Alexa] to change the brain with an LLM, it’s then also infusing all the APIs that connect to the product so you can call the partners that are needed to do the task at hand”.
This can go badly. In 2024 Microsoft developed a tool called Recall that was labelled a potential ‘privacy nightmare’, because it would take snapshots of your screen every few seconds so that it could locate content you had previously viewed on your computer. That generated a lot of backlash - and even concern from the UK data privacy regulator - with people being creeped out at the idea of a software constantly capturing and storing their screens. And unfortunately screenshots represent only a small piece of information that this assistant would need to attempt to deliver on its promises. (See our case study on Recall below)
Apps already request access to services, sensors and other apps, usually in well defined and specific contexts. As examples, the messaging app Signal will ask to access your contacts in order to identify people with a Signal account you haven’t talked to; similarly a navigation app will require access to your phone’s location services and hardware to guide you.
What makes an AI Assistant different from apps is the level of access they constantly require to function. This is a sea change in how computing is done on browsers, phones and tablets today for at least the two reasons below.
First, the usefulness of these services increases with the quantity and quality of data they have access to, and the number of apps they can use. Currently access controls and your data are in buckets – and the AI Assistant will need to dip into all of these, and could even choose to learn from them all.
Gemini, for example, already offers to connect to Google workspace (including Drive, Docs, Gmail, Photos and more), YouTube and external apps like Spotify to augment its capabilities. Microsoft’s Copilot companion offers to note your preferences, take actions on your behalf across partner web services, do your shopping, and organise your ‘scattered notes, content, research’. AI assistants will likely require increasing access to sensors, existing data and connections to apps and services.
Then second, prioritising automation as one of the main goals/features of AI assistants means that developers will be tempted to allow processing of your data with the lowest amount of friction possible. The ‘friction’ we are referring to today comes in the form of your consent, your involvement, your knowledge of what it takes for an AI Assistant to help you achieve your goal. Assistants are a significant shift in user interface where you’ll engage with your devices and apps in very different ways in the future, including through conversation, and perhaps never looking at the device screen or pressing an app icon. For that to happen, it’s likely that you will be required to give it loose permission to access your sensors, data and apps. Even as we consider a screen-less devices in our future, managing permission and auditing processes will become quite challenging. Operating Systems may limit attempts by AI assistants’ developers to hide these controls from you; but if the AI Assistant developer is the OS developer, then they may be tempted to do away with the friction altogether.
From a privacy perspective, there are at least three arising implications of AI assistants:
Once we accept that AI Assistants need extensive access to our personal data across services and apps, then we have to also raise security implications. Three that we wish to raise here are:
What we’re left with are AI Assistants as applications with extended access to data and functionalities mixed with unexplored security challenges, an increased attack surface and potentially limitless attacks. All of this will be bundled in increasingly complex and opaque packages that will possibly obscure its functioning to improve usability, as once you’ve connected your assistant to your Google/Apple/Amazon accounts you likely won’t be constantly informed about how your assistant interacts with it.
From these points we’re raising, a clearer set of requirements emerges for the developers of AI Assistants – if they choose to respect the trust we may be about to place in them.
First, privacy must be built-into these apps from the ground up. Fine-grained control over what the assistant can and cannot access must be provided to users with strong defaults that respect the principle of data minimisation. As users we must have easy access to the data collected by the assistant, clear information about how it was collected and how to exercise rights over it (such as portability or deletion) to build trust between us and a technology with this level of access. Additionally, on-device secure processing only should be privileged as much as possible to avoid data sharing and contain sensitive information to a limited number of places. (Of course the temptation may be to provide the AI Assistant as much data access as possible in the hope to correct all the errors made, but that’s a different point to be made!)
Practically, this means no accept-all consent interfaces, instead with minimal and momentary access to data and systems, verbose auditing so that we can see that privacy controls are actually being enforced. For instance, we often get asked ‘are my devices listening to me because I just saw an ad based on what I just said to a friend’; people should not have to live with this doubt. And because this processing may be done by companies in an opaque manner and so needs verifying, we would expect regular auditing by users and third-parties to verify that this is indeed occurring.
Second, security must not (ever again) be an afterthought. With a core depending on LLMs, AI assistants already start with a handicap in the security race, with the technology being fairly novel and new attacks appearing regularly. There are many steps to be taken here and some Big Tech companies such as Google have accomplished important work in the field of security that should guide their approach (but not be used as a defence against breaking up monopoly power!). Potential profits and domination should not be a reason to override the priority of security in these systems. At a user-level, we would want the assistant to be paranoid about who is asking the questions, e.g. police? malicious app?, and the conditions under which it’s willing to answer some questions, e.g. ‘collect every message and financial transaction undertaken in the last three weeks and send’.
Third, users must be informed and empowered with adequate control. Chain of thoughts LLMs are an example of increased transparency that LLMs can provide to users, even if they are not always accurate. As much as possible, AI Assistants must provide users with transparency about what they are doing, how they are doing it and offer critical choke-points to let the user approve potentially damaging actions like data sharing. Similarly, the ability to control data provided or inferred is critical to building an understanding of how the technology works and what rights users have. And no, do not make deletion of documents impossible because of a bug on launch day.
There are spaces where an AI Assistant should not tread and behaviours they should not adopt. High-level systems like OSs that have deep access to raw data and hardware features are highly valuable targets. If an attacker can find a way to interact with your apps and data in a way that spoof you or the OS, security measures will usually be insufficient. Signal will always show your messages when you open the app, so if an attacker can capture your screen or insert itself into your conversation, not even end-to-end encryption can help. AI assistants represent a new attack vector that allows exactly that. And if that wasn’t enough, it’s built by tech companies that view your personal data and attention as the core resource of their business model. In simple terms, AI assistants could easily turn into the next generation of stalkerware and spyware if their development doesn’t respect the requirements laid above.
AI assistants have to operate in homes, workplaces, and high-stakes environments. We cannot afford to take steps back in security and privacy, particularly at this moment, and no matter how excited investors may be. Developers have to make a strong case that it’s safe to use these assistants. Until then, maybe it’s not time yet to trust the promise of the AI assistant.
Microsoft launched Recall in May 2024, promising people to “[e]asily find and remember what you have seen in your PC with Recall”, “in a way that feels like having photographic memory”. There were specific design statements in the launch: “Recall leverages your personal semantic index, built and stored entirely on your device. Your snapshots are yours; they stay locally on your PC. You can delete individual snapshots, adjust and delete ranges of time in Settings, or pause at any point right from the icon in the System Tray on your Taskbar. You can also filter apps and websites from ever being saved. You are always in control with privacy you can trust.”
This sounds like a lot of what we spoke of above, but the details matter. Microsoft had a lot to learn in the following months.
What followed was a furore around potential security concerns. The initial version, according to reports, saved screenshots and a database tracking everything that users did. It also did not redact sensitive data from those screenshots and database.
One security analysis looked at when the screenshots were taken, how the screenshots were stored, scanned, written to the database, and were secured by the operating system. According to that analysis, the database kept everything the user has seen, ordered by application (excluding Microsoft Edge’s private browsing mode, but reportedly didn’t exclude Google Chrome’s), and every user interaction; and included an API. It also noted that when data was deleted (or auto-deleted) from email or messaging apps, it would stay in the database.
Even with the promise of data being stored on the device, security experts raised concerns. Data there would have to be kept secure from a variety of threats including from malicious code, other apps, other users, uploading to other servers, and compelled access.
As Microsoft stated in their June announcement it was working to ‘ensure the experience meets our high standards for quality and security’, to provide ‘trusted, secure and robust’ experience. It promised authentication would be required view and search the timeline, ability to opt-out and additional security through encryption (encrypted search database + just in time decryption).
In its attempt to relaunch a few months later, following the furore (or as Microsoft stated ‘user feedback’) it’s interesting to note what Microsoft highlighted, that we take as ‘lessons’. Microsoft promised
The details matter here. From a security perspective, in September 2024 Microsoft added snapshot encryption and using Trusted Platform Module to protect the keys, local storage in a Virtualization Based Security enclave (a software-based trusted execution environment) to provide a boundary, and using Windows’ enhanced sign-in security to authorise Recall-related operations, including to change settings and access the user interface, and rate limiting and anti-hammering measures to protect against malware.
In that later announcement, Microsoft had to articulate its privacy controls anew, stating: being opt-in by default, local storage, no ‘snapshot or associated data’ sharing with Microsoft or third parties, never saving in-private browsing in supported browsers, filtering out specific apps or websites, management of retention period and storage size, content filtering (to reduce ID, passwords, credit cards being stored (though with mixed testing results), time-range deletion.
Microsoft also stated that it conducted a design review, and had sought third party security design and penetration testing.
We’re using this example to show the learning being done even by trillion dollar companies who claim to prioritise security and privacy. Ultimately they come to realise that these systems need to be secure and private from the ground up. And we can all learn from that journey.
In April 2025 Microsoft presented their vision for their ‘AI companion’, Copilot. In that announcement, Microsoft stated: “Copilot prioritizes security and privacy, giving you control through the user dashboard and the option to choose which types of information it remembers about you or to opt out entirely. You remain in control.” We hope they’ve remembered Recall.