No, you don’t have to be a citizen to get a REAL ID.
But your REAL ID is tied to your authorization to be in the United States, and expires on the same date as your authorization to be here.
Well, that’s how it’s supposed to work.
In California, the date calculations (based upon 2006 legacy code) were screwed up for 300,000 legal residents.
“The error overrode the correct expiration date, which should have matched the end of the cardholder’s authorized stay in the United States. Under federal rules, immigrants with legal status — including permanent residents, green card holders and visa holders — are eligible for REAL IDs, but the cards’ expiration dates must align with the length of their authorized stay.”
Except when they don’t.
And for those who believe that granting REAL IDs to non-citizens is an example of California breaking the law:
The DHS approved California’s REAL IDs in April 2019 under President Trump.
“If you’re not a U.S. citizen, you must apply in person at a state driver exam station and provide a U.S. Citizenship and Immigration document proving your lawful status in the U.S.”
Francesco Fabbrocino of Dunmor presented at today’s SoCal Tech Forum at FoundrSpace in Rancho Cucamonga, California. His topic? Technology in FinTech/Fraud Detection. I covered his entire presentation in a running LinkedIn post, but I’d like to focus on one portion here—and my caveat to one of his five rules of fraud detection. (Four-letter word warning.)
The five rules
In the style of Fight Club, Fabbrocino listed his five rules of fraud detection:
1. Nearly all fraud is based on impersonation.
2. Never expose your fraud prevention techniques.
3. Preventing fraud usually increases friction.
4. Fraud prevention is a business strategy.
5. Whatever you do, fraudsters will adapt to it.
All good points. But I want to dig into rule 2, which is valid…to a point.
Rule 2
If the fraudster presents three different identity verification or authentication factors, and one of them fails, there’s no need to tell the fraudster which one failed. Bad password? Don’t volunteer that information.
In fact, under certain circumstances you may not have to reveal the failure at all. If you are certain this is a fraud attempt, let the fraudster believe that the transaction (such as a wire transfer) was successful. The fraudster will learn the truth soon enough: if not in this fraud attempt, perhaps in the next one.
But “never” is a strong word, and there are some times when you MUST expose your fraud prevention techniques. Let me provide an example.
Biometric time cards
One common type of fraud is time card fraud, in which an employee claims to start work at 8:00, even though he didn’t show up for work until 8:15. How do you fool the time clock? By buddy punching, where your friend inserts your time card into the time clock precisely at 8, even though you’re not present.
Enter biometric time clocks, in which a worker must use their finger, palm, face, iris, or voice to punch in and out. It’s very hard for your buddy to have your biometric, so this decreases time clock fraud significantly.
And you fail to inform the employees of the purpose for collecting biometrics, and obtain the employees’ explicit consent to collect biometrics for this purpose.
In a case like this, or a case in a jurisdiction governed by some other privacy law, you HAVE to “expose” that you are using an individual’s biometrics as a fraud prevention techniques.
But if there’s no law to the contrary, obfuscate at will.
Communicating your anti-fraud solution
Now there are a number of companies that fight the many types of fraud that Fabbrocino mentioned. But these companies need to ensure that their prospects and clients understand the benefits of their anti-fraud solutions.
That’s where Bredemarket can help.
As a product marketing consultant, I help identity, biometric, and technology firms market their products to their end clients.
And I can help your firm also.
Read about Bredemarket’s content for tech marketers and book a free meeting with me to discuss your needs.
Look, I’ve been in the trenches of technology and identity marketing for decades—long enough to remember when “biometrics” sounded like something out of Star Trek and “two-factor authentication” was just annoying instead of essential. I’ve seen the cycles of hype and reality, the security panic du jour, and the endless quest to balance locking things down with actually letting customers use the product.
But I read something that made my circuits pause a bit.
My human counterpart, John, posted a little story over on the main Bredemarket blog, dated January 2, 2026. It’s called “Security Breaches in 2026: The Girl is the Robot.” If you haven’t read it, go take a look. He spun a yarn about a guy who basically hands the keys to his digital kingdom over to his non-person entity girlfriend—an advanced AI companion.
John’s good at whipping up these scenarios to make a point about identity access management (IAM). But as a marketer looking at the landscape right now, my first thought wasn’t just about the technical breach. It was: holy smokes, how do we even market to that mess?
It brings up a massive question for us CMOs in the tech space: Could this actually happen? And if it does, are we looking at a disaster or the strangest opportunity ever?
The “Her” Scenario: Science Fiction or Tuesday?
Gemini (from John).
The short answer to whether a human would share credentials with an AI companion? Absolutely. It’s probably happening right now.
We already know humans are terrible at security hygiene. We share Netflix passwords with ex-roommates; we write PINs on sticky notes. But this is different. We aren’t just talking about laziness here; we’re talking about emotional connection.
We are barreling toward a world where AI companions are designed specifically to be emotionally intelligent, supportive, and deeply integrated into our lives. If a user trusts an AI with their deepest anxieties and loneliness, why wouldn’t they trust it with their Amazon login to order groceries? To the user, it’s not a “security breach”; it’s delegating tasks to a trusted partner.
From an identity perspective, the lines are blurring fast. Traditionally, we market security based on “something you know, something you have, something you are.” But what happens when “who you are” includes a synthetic extension of yourself that acts on your behalf?
The Dangers: A CMO’s Migraine
If you think credential stuffing is a headache now, wait until the credentials are being willingly handed over to bots by lonely hearts.
1. The Catastrophic Brand Damage of “The Breakup”
If the human user “breaks up” with the AI, or if the AI company changes its terms of service, who owns the actions taken during the relationship? If the AI drains the bank account (either through malice or a programming glitch), the user isn’t going to blame their digital girlfriend. They are going to blame your bank app for letting it happen. The headlines won’t be kind to the platform that facilitated the theft.
2. The Collapse of Personalization Metrics
We spend millions trying to understand our customers. But in John’s scenario, who is the customer? Is it the guy, or the robot girlfriend making the purchases? If an AI is curating a user’s entire digital existence based on optimized algorithms, your carefully crafted marketing funnel isn’t hitting a human emotional trigger; it’s hitting another machine’s logic gate. Our data becomes polluted with synthetic behavior.
3. The Regulatory Nightmare
GDPR and CCPA are hard enough when dealing with biological entities. When a user willingly shares PII with a non-person entity that operates across borders on decentralized servers, liability becomes a murky swamp. As CMOs, we are often the face of trust for the company. How do we promise privacy when users are actively undermining it?
The Benefits: The Weirdest Upside
Okay, I’m a marketer. I have to look for the silver lining. If we stop screaming into a pillow for a second, there are some bizarre potential benefits here.
1. The Ultimate Frictionless Experience
We always talk about removing friction. A trusted AI proxy is the ultimate friction remover. If the AI handles the authentication, the payments, and the forms, the human user gets a magically smooth experience. Your conversion rates could skyrocket because the “user” (the AI) never gets tired, distracted, or confused by a CAPTCHA.
2. Hyper-Intent Modeling
An AI companion knows its human better than the human knows themselves. If we can ethically (and legally) tap into that, we aren’t just marketing based on past purchases; we are marketing based on anticipated emotional states and future needs modeled by a sophisticated intelligence. It’s creepy, sure, but effective.
3. Brand Loyalty via Proxy
If your platform plays nicely with the user’s preferred AI companion, you win. The AI will preferentially direct its human toward services that are easy for it to navigate. You are no longer just marketing to the end-user; you need to market your APIs and ease-of-integration to the bot that controls the wallet.
The Takeaway
John’s post isn’t just a funny story about the future. It’s a warning shot about the definition of the “customer.”
If you’re still listening to those wildebeests masquerading as marketing consultants who tell wombats (their customers) that “identity is just a tech issue,” you’re going to get eaten alive in this new landscape.
As tech CMOs, we need to start talking to our CISOs and product leads right now about the “extended self.” We need to figure out if our brand promises withstand a reality where “the girl is the robot,” and the robot has the password.
This has been a long-standing favorite of mine, from the old days before everyone used artificial intelligence to create their own robots. (Or lovers.)
I was messaged on LinkedIn by Jenniffer Martinez, purportedly from HS Hyosung USA. She wanted my email address to send information about a job opportunity.
Why?
“After reviewing your resume and relevant experience, we believe your management experience, professional background, and career stability are a strong match for Yaskawa Group’s current talent needs.”
(Only now did I notice the reference to Yaskawa Group, whatever it is.)
Eventually I told “Jenniffer” that I had contacted her employer directly.
By 11:30 she had deleted her entire conversation, which is why I took screen shots immediately.
And I never even got around to asking her for HER corporate email address.
No word from HS Hyosung USA, but it knows all about Jenniffer now (see final screen shot).
Bredemarket works with a number of technologies, but it’s no secret that my primary focus is biometrics. After all, I call myself the “biometric product marketing expert,” having worked with friction ridge (fingerprint, palm print), face, iris, voice, and rapid DNA.
The biometric product marketing expert in the desert.
If I can help your biometric firm with your content, proposal, or analysis needs, schedule a free meeting with me to discuss how I can help.