Connect with us


Podcast: Attention shoppers–you’re being tracked



In some stores, sophisticated systems are tracking customers in almost every imaginable way, from recognizing their faces to gauging their age, their mood, and virtually gussying them up with makeup. The systems rarely ask for people’s permission, and for the most part they don’t have to. In our season 1 finale, we look at the explosion of AI and face recognition technologies in retail spaces, and what it means for the future of shopping.

We meet:

  • RetailNext CTO Arun Nair
  • L’Oreal’s Technology Incubator Global VP Guive Balooch
  • Modiface CEO Parham Aarabi
  • Biometrics pioneer and Chairman of ID4Africa Joseph Atick


This episode was reported and produced by Jennifer Strong, Anthony Green, Tate Ryan-Mosley, Emma Cillekens and Karen Hao. We’re edited by Michael Reilly and Gideon Lichfield.



Strong: Retailers have been using face recognition and AI tracking technologies for years.

[Audio from Face First: What if you could stop retail crime before it happens by knowing the moment a shoplifter enters your store? And what if you could know about the presence of violent criminals before they act? With Face First you can stop crime before it starts.]

Strong: That’s one of the largest providers of this tech to retail stores. It detects faces, voices, objects and claims it can analyze behavior. But face recognition systems have a well-documented history of misidentifying women and people of color. 

[Sound from 2019 Congressional hearing on facial recognition (Ocasio-Cortez): We have a technology that was created and designed by one demographic that is only mostly effective on that one demographic. And they’re trying to sell it and impose it on the entirety of the country?]

Strong: This is Representative Alexandria Ocasio-Cortez at a 2019 congressional hearing on facial recognition. Photo technologies work better on lighter skin. And datasets used by companies to train facial analysis systems are largely based on faces collected from the internet where content tends to skew white, male and western. 

[Sound from 2019 Congressional hearing on facial recognition (Ocasio-Cortez): And do you think that this could exacerbate the already egregious, uh, inequalities in our, in our criminal justice system]

[Sound from 2019 Congressional hearing on facial recognition (Buolamwini): And It already is.]

Strong: Joy Buolamwini is an activist and computer scientist.

[Sound from 2019 Congressional hearing on facial recognition (Buolamwini): So, there’s a case with Mr. Bah, an 18-year-old African American man who was misidentified in Apple stores as a thief. And in fact, he was falsely arrested multiple times because of this kind of misidentification.

Strong: As awareness of these issues grows, more places are looking to put restrictions around its use such as in Portland, Oregon, which recently passed the most sweeping ban on face ID in the US.

[Sound from store in Portland, Oregon: please look into the camera for entry]

Strong: The ban takes effect in January and when it does that voice and camera will go away from places like this food store where the tech unlocks the door to late night shoppers. But use elsewhere is moving well beyond fighting crime (and is starting to play other retail roles) like remembering your past orders and payment details.

Miller: These face-based technologies, uhh artificial intelligence, machine vision allow us to see our customer in the offline world like amazon sees its customer in the online world. That allows us to create tailored experiences for the customer and also allows us to directly target that customer in new ways when they come back to the restaurant.

Strong: That’s the chairman of Cali Group, John Miller, its fast-food restaurant Caliburger tries out technologies it later markets to the entire industry. Other retailers use face recognition to know when VIP shoppers or celebrities are in their stores, not unlike this scene from the film Minority Report where as Tom Cruise strolls through a mall, his eyes are scanned and the ads address his character by name.

[Sound from Minority Report where voices address John Anderson in person]

Strong: The face measurements powering these applications can also be used for many other things besides just identifying someone. For example, some shopping malls use it to help set their store rents by counting how many people walk by, and using face data to gauge gender, age, and other demographics. Sometimes face recognition cameras are even hidden inside in mall directories. And inside stores, retailers use it to better understand what shoppers are interested in. It’s also embedded within shopping apps and store mirrors that let people try on anything from eyeglasses to makeup virtually. 

I’m Jennifer Strong and this episode, we wrap up our first season (and our latest miniseries on face recognition) with a look at how it’s used to watch, understand and influence your shopping habits.


Strong: So I’m out front of what used to be the largest store in the world. This is Macy’s on 34th Street in Manhattan. The building fills an entire city block and in some ways it’s kind of the center of gravity for the holiday shopping season here as, among other things, the inspiration for one of New York’s most famous Christmas films, Miracle on 34th Street. 

But the company may also have a history of using face recognition and a lawsuit was filed about that in Illinois which has a biometric privacy law requiring companies get permission before using it on customers. That suit alleges Macy’s is a client of ClearviewAI. We’ve had the founder on this show Hoan Ton-That and his product works by matching images, in this case of shoppers or shoplifters, against a database of perhaps billions of photos taken from social media posted by people who haven’t changed their settings to make the photos private just to their friends.

Now, New York City’s councilmembers just passed a biometrics measure here that if signed by the mayor will make retailers here also tell shoppers that face recognition is being used and perhaps what’s happening with that data. But you know it’s too soon to say what that might look like. I mean does walking as part of a big crowd of shoppers past a wall plaque that says face recognition is present, does that equal being informed, let alone giving consent? But I’m going to go inside with my producer, Anthony Green, and see if we can find totally different applications of face mapping to show you.

Several of these beauty counters have iPads that double as mirrors with augmented reality. We tried out three of them just one though asked for consent to analyze our faces. Two of the systems saw us just fine through our masks. The other didn’t recognize our faces at all.

I walked up to a mirror and it says my lighting is okay. Come closer until your face fills a circle. Apparently I have dark circles, uneven texture. irritation and redness and eyelines. At least we’re on the less side? I don’t know. Woah. Hey Anthony, you should see this. I wasn’t sure it was doing anything and now look in the mirror. 

Green: Wow. 

Strong: Right?

Green: Wow.

Strong: I don’t really have words for describing this, but it is so funny seeing myself this made up. 

Green: Just kind of like glammed up.

Strong: Yeah. I’m like super glammed up. And literally all I was doing was looking in this mirror and then I looked down on an iPad and Holy, wow.

Green: This is working with your mask on.

Strong: This is with my mask on. And if I pull my mask down, I am made up everywhere. 

Green: Oh yea.

Strong: Like glossed and all. Oh, look at you. 

Green: Wow. 

Strong: Okay, so Anthony just took a step over towards me and now he’s made up to the nines. Okay. These experiences are among the many many ways that face mapping can be applied.

But because they’re so controversial most brands simply don’t want to talk about it. And mostly, they don’t have to. There’s no national requirement that companies disclose the way they gather or use our biometric data even though we can imagine a not-so-distant future when that data becomes more important than any document we have. This personal data is likely to replace all of them proving who we are and what we own.

Most of what we know about the use of face recognition by retailers started in 2013 when it became public that identity company NEC had about a dozen brands and hotels as clients and they were using its face-reading technology to identify celebrities and other VIPs as they walked through their doors. 

The following year Facebook announced it applied neural networks to face ID for the first time, making it work significantly better. And retailers, including Walmart, began testing it as a way to identify people caught shoplifting. 

By 2016 fast food companies were experimenting with other use cases. One partnership, between KFC and the Chinese tech giant Baidu, recommended menu items to customers based on their age and mood as deemed by face scanning. These days it’s also possible to pay with your face, though so far, these applications haven’t really caught on. And so, wherever you shop, it’s reasonable to assume you might encounter some aspect of this technology and it could be combined with any number of other trackers. But it’s equally true that much of the tracking that’s done in retail stores using computer vision involves no facial recognition at all. 

Nair: If you build a website today, there are a lot of tools available that you can use to give you data, like how many people visited your website, who they were, how they navigated your website and so on and for e-commerce sites the eventual purchase activity as well. And you can use all of this data to understand visitor behavior and optimize your site. We do the exact same thing, but for physical spaces. My name is Arun Nair. I’m the CTO and co-founder of RetailNext.

Strong: Their tracking software is deployed in offices, museums, even bowling alleys, but their primary market is retail. Ceiling cameras equipped with computer vision track customers as they journey through the store. It can guess basic demographic information like gender, who’s an employee—based on whether they go behind the register, even interactions between employees and customers. 

Nair: We even have a prediction algorithm that will tell you based on historical information when your store is going to be busy later in the day, later in the week. And it is extremely helpful for staffing. So making sure that when you do expect a peak, that there are people there to assist shoppers and they’re not standing in queue and so on as well as you’re not always staffed when no one needs to be there.

Strong: He says the company is capable of determining what you’re looking at, but it doesn’t track eye gaze, expressions, or faces. And they don’t individually identify anyone.  

Nair: We do not know who they are as individuals, and we specifically try not to as well. And in actually a lot of cases, once we get that information, we throw away the video or we blur the video.

Strong: When it comes to privacy, he believes systems using face recognition for identity should be opt-in

Nair: Consent is not just about like, Oh, I put my data out there so you can do what you want. I think consent is also about  you know, we want you to do this so that we can do this in return for you. Are you okay with that?

Strong: But he admits that’s easier said than done.

Nair: It’s not easy to opt out of those things. And even if you opt out, the challenge is that let’s say, you say that, Hey, I want to opt out of my face.   As a technology company, I still have to store a digitized version of your face to make sure I don’t track you again in the future cause next time I see your face, I need something to map against to say that, Oh, I should be dropping this person’s face. But then again, you know, in a weird way, I’m now storing a digitized version of your face, which. Again, it’s not really your face, but it is a representation of it. 

Strong: And these challenges aren’t going away. Most tracking technologies aren’t regulated, and we simply don’t know how often things like face data gets captured. What is clear the retail industry is shifting to a world that’s centered around real-time analysis of customer experiences. 

Nair: I think they’re going to see more and more of that moving forward, where there’s fewer purchases actually happening in these locations, but that’s kind of how you’re learning about the brand. [00:12:15] Almost like advertising, as well as kind of building a brand loyalty. 

Strong: Tracking customers and their interaction with the store doesn’t just help retailers know what’s selling  It also gives them insight on what customers want. 

Nair: You introduce a new product. And you want to make sure that people are seeing that product. Our algorithms will tell you if people actually go into an area of the store and interact with a product and actually make a purchase afterwards.

Balooch: I think that it’s a combination of AI with physical objects that creates really an exciting moment in time. You know, you could never really try a trend and then actually dispense it. That wasn’t possible ever. But now because of AI, we are able to really go through trends really quickly. We’re able to curate trends, we’re able to give people what they desire. My name is Guive Balooch and I run the global technology incubator at L’Oreal. I’ve been at the company for 15 years and my job is to find the intersection between beauty and technology.

Strong: L’Oreal is the world’s largest cosmetics company with Estee Lauder, Maybelline, Garnier and countless other consumer brands under its corporate umbrella. 

Balooch: We started about eight years ago with an augmented reality app called makeup genius. That was the world’s first virtual try-on. And since then we’ve launched projects around personalized beauty like skincare personalization, foundation personalization. We’ve launched a UV sensor at the Apple store that’s a wearable that has no battery and can measure your UV exposure. And now we’re, we’re moving more and more towards mass personalization and finding ways to combine technologies like AR and AI to create new physical objects that can be magical for beauty consumers and hopefully delight our users.

Strong: And this is harder than it might sound. Designing experiences that let customers try on makeup in augmented reality presents huge technical challenges for face detection.

Balooch: You need to detect where the eye is and where the eyebrow is. And it has to be at a level of accuracy that when the product’s on there, it doesn’t look like it’s not exactly on your lip. And it’s, it’s funny because I come from an academic background with a PhD. So I didn’t realize how complicated that specific part of this technology is. I thought, “Oh, it’s okay. We’ll just get the software. It will be easy. We’ll just make it work.” But it turns out no, it’s really complicated because people’s lips can vary in shape, the color between your skin tone and your lip can also be very different. And so you need to have an algorithm that can detect it and make sure it works on people from very light to very dark skin. 

Strong: And he says one of the largest impacts of AI in the beauty market could be more inclusivity—something the industry has long struggled with.

Balooch: I’m under this, you know, strong belief that inclusivity is the future of beauty and inclusivity means that every human being has the right to have a product that is what they need for themselves and to showcase to the world how they want to be showcased. And I think that only through things like AI and tech, will we be able to reach that level of personal relationship with people’s desires for their beauty habits.

Strong: Those habits are shaped around our skin. And skin tone has historically been one of the hardest technical and cultural challenges.

Balooch: We launched this project called which is this foundation blender. And when I first started this project, I thought it was going to be very simple because when I went to Home Depot umm I’m not really a handyman, but I went with my, my dad a lot to Home Depot and he would buy paint. He would match the paint and they would just make the paint right there. And I said, okay, it’s that easy? So when we first started the project, we realized, okay, you know, you just take a skin tone from a piece of, you know, a paper and you can just match the foundation. And I realized later that our skin is not like a wall, it’s biological tissue that changes depending on what kind of skin tone you have.

Strong: In short, the algorithm didn’t work. 

Balooch: And so we had to stop and spend another six months to improve it. First we did that with a little device that kind of measures your skin tone, using a physical object, because your skin tone is hard to measure if you don’t actually touch the skin cause the light can change the color of your skin. And so depending on if you’re outside or if you’re inside, you could have a big difference in the measurement. But not anymore. Thanks to AI, I think more and more with AI, we’re going to be able to get accurate measurements. We have to test them and make sure that they work as well as objects. But once we get to a point, when we think we’re getting close to that, then you can solve some really, really big challenges. And in foundation, 50% of women can’t find the right shade of foundation. And there’s no way that the number of products on the shelf will ever solve that because you will always have more skin tones in the world than products you can put on the shelf.

Strong: And the future could open up a whole new class of personalized beauty tools.  

Balooch: We can make objects that are, you know, not huge–handheld–and can do incredible things. Like in the future, you could imagine that you can dispense eyeshadow on your eyelid automatically just through detecting the face and being able to have an object that could dispense it. 

Strong: To build that future, L’Oreal acquired a company called Modiface which makes augmented reality tools for more than 70 of the world’s top beauty brands.

Aarabi: One big step that happened a few years ago was going from photos to live video simulation. Really hard feat technologically, but really impactful on the consumer experience. Instead of having to take a photo and upload it, they could see a live video. 

Strong: Parham Aarabi is the Founder and CEO of Modiface. 

Aarabi: The next big step that I see that I’m really excited about is a combination of AI understanding of the face, along with our simulation. So not only telling you, okay, so you choose a lipstick and this is what it looks like, but saying, because you chose this lipstick and because your, you know, you have blue eyes, we believe this eye shadow might match it the best.

Strong: His background is in face and lip tracking.

Aarabi: And so we had created this sample demo where you could track someone’s lips and swap the lips with a celebrity, for example. My co-founder had the idea that before we do this, we should actually apply some changes on the, on the skin. And so it was really the combination of these two ideas that became the foundation of Modiface. 

Strong: The beauty industry thrives on the in-person shopping experience. And even though e-commerce sales have long been on the rise this sector has been a lot slower than others. For context, the top ecommerce seller in beauty of 2018 was shampoo. But the pandemic is speeding things up. Online sales at beauty giant Sephora jumped 30 percent in the U.S. this year. And it’s also partnered with Modiface to develop an app that acts as a virtual store, complete with product tutorials and an augmented reality beauty counter. 

Aarabi: You see a try-on button, you press that, and a window opens up. You see your own video in that window, but with different virtual products being shown.

Strong: And building consumer trust in these simulated products means engineering an experience as seamless as looking in a mirror. 

Aarabi: If someone actually tries on a lipstick and a hair color and then videotapes themselves versus using our technology and then having a virtual simulation of those products, the two should be indistinguishable. The lag, within the simulation being applied versus when you’re looking at your face and you’re seeing movements needs to be not apparent to the user. And so these are huge challenges. One is of realism. You don’t want the eyeliner to be flickering on someone’s eyes and the second is to do it so fast that on a website in live video, you don’t notice any lag. So these are major, major challenges.

Strong: And it’s more than just cosmetics. Elements of face detection are increasingly used in medicine to diagnose disease. And he believes in future their products will detect all kinds of skin disorders. 

Aarabi: So we’ve been pushing on this skin assessment, um, direction by looking at someone’s image. And based on that, knowing what skin care products are best for them, and more, the more we do this and the more that better we train our AI systems, we find that they’re increasing in the level of accuracy matching that of dermatologists. And I think if you follow that line, that this AI, that can actually not replace dermatologists, but really helped them as.. an objective tool that can look at someone’s face and make recommendations.

Strong: It feels like there’s more awareness of face recognition of its risks, immaturies and biases but also its increased presence in our lives and just raw potential. To me, it seems like we’ve just scratched the surface – in this messy digital race to something different and big. And it got me wondering how might one of its inventors feel about all this?

Atick: I started working on the human brain about a year after I graduated and made together with my, collaborators made some fundamental breakthroughs, which led to the creation of a field called the biometric industry and the first commercially viable face recognition. That’s why people refer to me as a founding father of face recognition and the biometric industry.

Strong: That’s Dr. Joseph Atick. He developed one of the first face recognition algorithms back in 1994.

Atick: The algorithm for how a human brain would recognize familiar faces became clear while we’re doing mathematical research at the Institute for advanced study in Princeton.

Strong: But the technology needed to capture those faces wasn’t yet in everyone’s pockets. 

Atick: At the time, computers did not have cameras. Phones that had cameras did not exist. We had to build the eyes for the brain. We had a brain, we thought we knew how the brain would analyze signals, but we did not have the eyes that would get the information and the visual signal to the brain.

Strong: Webcams came along in the 90s and computers with video capabilities arrived on the market a few years after.

Atick: And that was an exciting time because all of a sudden the brain that we had built had finally the pair of eyes that would be necessary to, to see.

Strong: This was the breakthrough he and his team needed to bring their concept to life. So they started coding.

Atick: it was a long period of months of programming and failure and programming and failure

Strong: But eventually…

Atick: And one night, early morning, actually, we had just finalized, um, a version of the algorithm. We submitted the, source code for compilation in order to get a run code. And we stepped out, I stepped out to go to the washroom. And then when I stepped back into the room it spotted my face, extracted it from the background and it pronounced “I see Joseph”. And that was the moment where the hair on the back–I felt like something had happened. We were a witness. And I started, um, to call on the other people who were still in the lab and each one of them, they would come into the room. And I would say, it would say, I see Norman. I would see Paul, I would see Joseph. And we would sort of take turns running around the room just to see how many it can spot in the room.

Strong: They had built something that had never been built before. Months of math and coding and long nights seemed to be paying off. But within a few years that excitement turned to concern.

Atick: My, my concern about the technology that I helped create and invent started very quickly after I had invented it. I saw a future where our privacy would be at jeopardy if we did not put in place protection measures to prevent the abuse of this powerful technology.

Strong: And he wanted to do something about it.

Atick: So in 1998, I lobbied the industry and I said, we need to put together principles for responsible use. And this is where an organization called IBIA was born in 1998 as an industry association to promote responsible use. Um, and so I was the founder of that, that organization. And I felt good for a while because I felt we have gotten it right. I felt we’ve invented the technology, but then we put in place a responsible use code to be followed by whatever is the implementation. However, that code did not live the test of time. And the reason behind it is we did not anticipate the emergence of social media.

Strong: Face recognition relies on a database of images. The size, quality, and privacy conditions of this database is largely what determines how safe or intrusive the technology is. In 1998, Atick built his databases by manually scanning thousands of pictures and tagging them with names. It was tedious and limiting in size.

Atick: We have allowed the beast out of the bag by feeding it billions of faces and helping it by tagging ourselves. We are now in a world where machine learning is now allowing for the emergence of over 400 different algorithms of face recognition in the world. Therefore, any hope of controlling and requiring everybody to be, to be responsible in their use of face recognition is difficult.

Strong: And this is made worse by scraping, where a database is created by scanning the entire internet for public photos.

Atick: And so I began to panic in 2011, and I wrote an op-ed article saying it is time to press the panic button because the world is heading in a direction where face recognition is going to be omnipresent and faces are going to be everywhere available in, in, in databases. Computing power is becoming very, very massive to the point that we could potentially recognize billions of people. And at the time people said I was an alarmist, but they’re realizing that it’s exactly what’s happening today.

Strong: So in a way, he’s kind of lobbying against his own invention even though he still uses biometrics to help build things he believes might benefit the greater good like digital ID for people in developing nations.

Atick: The chilling effect is something that is unforgivable. If I cannot go outside in the street, because I believe somebody’s using an iPhone, could take a picture of me and connect me to my online profile and, this online and offline connection is, is a dangerous thing. And it’s happening right now.

Strong: And he thinks we urgently need some legal ground rules.

Atick: And so it’s no longer a technological issue. We cannot contain this powerful technology through technology. There has to be some sort of legal frameworks.

Strong: The way he sees it, the technological edge will keep pushing forward—with AI at the forefront. But the people building and using it? They’re at the center.

Atick: I believe there has to be some harmony between what technology can do for us and helps us live with dignity and have easier lives and connect with the people we love, but at the same time, it has to be within what our morals and our expectations as human beings allow it to be. 

Strong: In other words, once again… it seems up to us. This episode was reported and produced by me, Anthony Green, Emma Cillekens, Tate Ryan-Mosley and Karen Hao. We’re edited by Michael Reilly and Gideon Lichfield. Thanks too to Kate Kaye with the Banned in PDX podcast. That’s it for season one. Thanks so much for choosing to spend your time with us. We’ll meet you back here in the new year until then happy holidays and… Thanks for listening, I’m Jennifer Strong. 


Lyron Foster is a Hawaii based African American Musician, Author, Actor, Blogger, Filmmaker, Philanthropist and Multinational Serial Tech Entrepreneur.

Continue Reading


Hard-coded key vulnerability in Logix PLCs has severity score of 10 out of 10



Hard-coded key vulnerability in Logix PLCs has severity score of 10 out of 10

Enlarge (credit: Rockwell Automation)

Hardware that is widely used to control equipment in factories and other industrial settings can be remotely commandeered by exploiting a newly disclosed vulnerability that has a severity score of 10 out of 10.

The vulnerability is found in programmable logic controllers from Rockwell Automation that are marketed under the Logix brand. These devices, which range from the size of a small toaster to a large bread box or even bigger, help control equipment and processes on assembly lines and in other manufacturing environments. Engineers program the PLCs using Rockwell software called Studio 5000 Logix Designer.

On Thursday, the US Cybersecurity & Infrastructure Security Administration warned of a critical vulnerability that could allow hackers to remotely connect to Logix controllers and from there alter their configuration or application code. The vulnerability requires a low skill level to be exploited, CISA said.

Read 9 remaining paragraphs | Comments

Continue Reading


EC roundup: BNPL startups, growth marketing tips, solid state battery market map, more



When I needed a new sofa several months ago, I was pleased to find a buy now, pay later (BNPL) option during the checkout process. I had prepared myself to make a major financial outlay, but the service fees were well worth the convenience of deferring the entire payment.

Coincidentally, I was siting on said sofa this morning and considering that transaction when Alex Wilhelm submitted a column that compared recent earnings for three BNPL providers: Afterpay, Affirm and Klarna.

I asked him why he decided to dig into the sector with such gusto.

Full Extra Crunch articles are only available to members.
Use discount code ECFriday to save 20% off a one- or two-year subscription.

“What struck me about the concept was that we had just seen earnings from Affirm,” he said. “So we had three BNPL players with known earnings, and I had just covered a startup funding round in the space.”

“Toss in some obvious audience interest, and it was an easy choice to write the piece. Now the question is whether I did a good job and people find value in it.”

Thanks very much for reading Extra Crunch this week! Have a great weekend.

Walter Thompson
Senior Editor, TechCrunch

As BNPL startups raise, a look at Klarna, Affirm and Afterpay earnings

Pilot CEO Waseem Daher tears down his company’s $60M Series C pitch deck

Smashing brick work with hammer

Image Credits: Colin Hawkins (opens in a new window) / Getty Images

I avoid running Extra Crunch stories that focus on best practices; you can find those anywhere. Instead, we look for “here’s what worked for me” articles that give readers actionable insights.

That’s a much better use of your time and ours.

With that ethos in mind, Lucas Matney interviewed Pilot CEO Waseem Daher to deconstruct the pitch deck that helped his company land a $60M Series C round.

“If the Series A was about, ‘Do you have the right ingredients to make this work?’ then the Series B is about, ‘Is this actually working?’” Daher tells TechCrunch.

“And then the Series C is more, ‘Well, show me that the core business is really working and that you have unlocked real drivers to allow the business to continue growing.’”

Can solid state batteries power up for the next generation of EVs?


Image Credits: Bryce Durbin

A global survey of automobile owners found three hurdles to overcome before consumers will widely embrace electric vehicles:

  • 30-minute charging time
  • 300-mile range
  • $36,000 maximum cost

“Theoretically, solid state batteries (SSB) could deliver all three,” but for now, lithium-ion batteries are the go-to for most EVs (along with laptops and phones).

In our latest market map, we’ve plotted the new and established players in the SSB sector and listed many of the investors who are backing them.

Although SSBs are years away from mass production, “we are on the cusp of some pretty incredible discoveries using major improvements in computational science and machine learning algorithms to accelerate that process,” says SSB startup founder Amy Prieto.


Dear Sophie: Which immigration options are the fastest?

lone figure at entrance to maze hedge that has an American flag at the center

Image Credits: Bryce Durbin/TechCrunch

Dear Sophie:

Help! Our startup needs to hire 50 engineers in artificial intelligence and related fields ASAP. Which visa and green card options are the quickest to get for top immigrant engineers?

And will Biden’s new immigration bill help us?

— Mesmerized in Menlo Park


Why F5 spent $2.2B on 3 companies to focus on cloud native applications

Dark servers data center room with computers and storage systems

Image Credits: Jasmin Merdan / Getty Images

Founded in 1996, F5 has repositioned itself in the networking market several times in its history. In the last two years, however, it spent $2.2 billion to acquire Shape Security, Volterra and NGINX.

“As large organizations age, they often need to pivot to stay relevant, and I wanted to explore one of these transformational shifts,” said enterprise reporter Ron Miller.

“I spoke to the CEO of F5 to find out the strategy behind his company’s pivot and how he leveraged three acquisitions to push his organization in a new direction.”


DigitalOcean’s IPO filing shows a two-class cloud market

Cloud online storage technology concept. Big data data information exchange available. Magnifying glass with analytics data

Image Credits: Who_I_am (opens in a new window) / Getty Images

Cloud hosting company DigitalOcean filed to go public this week, so Ron Miller and Alex Wilhelm unpacked its financials.

“AWS and Microsoft Azure will not be losing too much sleep worrying about DigitalOcean, but it is not trying to compete head-on with them across the full spectrum of cloud infrastructure services,” said John Dinsdale, chief analyst and research director at Synergy Research.


Oscar Health’s initial IPO price is so high, it makes me want to swear

I asked Alex Wilhelm to dial back the profanity he used to describe Oscar Health’s proposed valuation, but perhaps I was too conservative.

In March 2018, the insurtech unicorn was valued at around $3.2 billion. Today, with the company aiming to debut at $32 to $34 per share, its fully diluted valuation is closer to $7.7 billion.

“The clear takeaway from the first Oscar Health IPO pricing interval is that public investors have lost their minds,” says Alex.

His advice for companies considering an IPO? “Go public now.”


If Coinbase is worth $100 billion, what’s a fair valuation for Stripe?

Last week, Alex wrote about how cryptocurrency trading platform Coinbase was being valued at $77 billion in the private markets.

As of Monday, “it’s now $100 billion, per Axios’ reporting.”

He reviewed Coinbase’s performance from 2019 through the end of Q3 2020 “to decide whether Coinbase at $100 billion makes no sense, a little sense or perfect sense.”


Winning enterprise sales teams know how to persuade the Chief Objection Officer

woman hand stop sign on brick wall background

Image Credits: Alla Aramyan (opens in a new window) / Getty Images

A skilled software sales team devotes a lot of resources to pinpointing potential customers.

Poring through LinkedIn and reviewing past speaker lists at industry conferences are good places to find decision-makers, for example.

Despite this detective work, GGV Capital investor Oren Yunger says sales teams still need to identify the deal-blockers who can spike a deal with a single email.

“I call this person the Chief Objection Officer.


3 strategies for elevating brand authority in 2021

Young woman standing on top of tall green bar graph against white background

Image Credits: Klaus Vedfelt / Getty Images

Every startup wants to raise its profile, but for many early-stage companies, marketing budgets are too small to make a meaningful difference.

Providing real value through content is an excellent way to build authority in the short and long term,” says Amanda Milligan, marketing director at growth agency Fractl.


RIBS: The messaging framework for every company and product

Grilled pork ribs with barbecue sauce on wooden background

Image Credits: luchezar (opens in a new window) / Getty Images

The most effective marketing uses good storytelling, not persuasion.

According to Caryn Marooney, general partner at Coatue Management, every compelling story is relevant, inevitable, believable and simple.

“Behind most successful companies is a story that checks every one of those boxes,” says Marooney, but “this is a central challenge for every startup.”


Ironclad’s Jason Boehmig: The objective of pricing is to become less wrong over time

On a recent episode of Extra Crunch Live, Ironclad founder and CEO Jason Boehmig and Accel partner Steve Loughlin discussed the pitch that brought them together almost four years ago.

Since that $8 million Series A, Loughlin joined Ironclad’s board. “Both agree that the work they put in up front had paid off” when it comes to how well they work together, says Jordan Crook.

“We’ve always been up front about the fact that we consider the board a part of the company,” said Boehmig.

TC Early Stage: The premiere how-to event for startup entrepreneurs and investors

From April 1-2, some of the most successful founders and VCs will explain how they build their businesses, raise money and manage their portfolios.

At TC Early Stage, we’ll cover topics like recruiting, sales, legal, PR, marketing and brand building. Each session includes ample time for audience questions and discussion.

Use discount code ECNEWSLETTER to take 20% off the cost of your TC Early Stage ticket!

Continue Reading


With $62.5M in debt financing, Road Runner Media puts digital ads behind commercial vehicles



If Southern California-based Road Runner Media succeeds, you’ll start seeing a lot more ads while you’re driving.

That’s because the startup is placing digital screens on the back of technicians’ vans, delivery vehicles, buses and other commercial vehicles. Those screens can show both ads and serve as a brake light — according to founder and chairman Randall Lanham, the brake light functionality is required if you’re putting a sign on the back of a vehicle.

“The way we look at it, we are a digital brake light,” Lanham said. Yes, the brake light is showing ads, but “the driver touching the brakes interrupts the ad.” (The sign can also indicate turns, reversing and emergency flashers. You can see a mock-up ad in the image above, and real footage in the video below.)

To pursue this idea, Lanham (who described himself as a “recovering attorney”) enlisted Chris Riley as CEO — Riley’s experience includes several years as CEO of PepsiCo Australia and New Zealand. And the company announced this week that it has secured $62.5 million in debt financing from Baseline Growth Capital.

The idea of putting ads on moving vehicles isn’t new. There are, of course, ads on the tops of taxis, and startups like Firefly are also putting digital signage on top of Ubers and Lyfts. But Riley said Road Runner’s ruggedized, high-resolution LCD screens are very different, due to their size, quality and placement.

“[Taxi-top ads] don’t have the color, the brilliance, the clarity,” he said. “We can run a true video ad on the screen.”

Riley also said the ads can be targeted based on GPS and time of day, and that the company eventually plans to add sensors to collect data on who’s actually seeing the ads.

As for concerns that these big, bright screens might distract drivers, Lanham argued they’re actually attracting driver’s eyes to exactly where they should be, and creating a brake light that’s much harder to ignore.

“Your eyes are affixed on the horizon, which is what the [Department of Transportation] wants — as opposed to on the floor or the radio or directly off to the left or right,” he said. “That’s where your safest driving occurs, when your eyes are up above the dashboard.”

In fact, Lanham said he’s “very passionate” about the company’s mission, which in his view will make roads safer, and is creating a platform that could also be used to spread public service messages.

“We have the ability to retrofit any vehicle and make it safer on the highways,” he added. “I really, truly believe that we will save lives, if we already haven’t.”

The company says it already has 150 screens live in Atlanta, Boulder, Chicago, Dallas and Los Angeles, with plans to launch screens in Philadelphia and Washington, D.C. in March.


Continue Reading