Lensa AI, the viral app making ‘magic avatars,’ raises red flags for artists

If your Instagram is awash in algorithmically generated portraits of your friends, you aren’t alone. After adding a new avatar generation tool based on Stable Diffusion, the photo editing app Lensa AI went viral over the last few days, with users sharing their uncanny AI-crafted avatars (and the horrible misfires) in stories and posts.

Lensa’s fun, eminently shareable avatars mark the first time that many people have interacted with a generative AI tool. In Lensa’s case, it’s also the first time they’ve paid for computer-generated art.

Stable Diffusion itself is free and a lot of people are playing around with it for research purposes or just for fun. But Lensa and other services like it — Avatar AI and Profilepicture.AI, to name a few — are making money by selling the computing cycles required to run the prompts and spit out a set of images. That certainly changes the equation a little.

Lensa is built on Stable Diffusion’s free, open source image generator but acts as a middleman. Send Lensa 10-20 selfies and $7.99 ($3.99 if you sign up for a free trial) and the app does the heavy lifting for you behind the scenes, handing back a set of stylized portraits in an array of styles like sci-fi, fantasy and anime. Anyone with sufficient processing power can install Stable Diffusion on a machine, download some models and get similar results, but Lensa’s avatars are impressive and Instagram-ready enough that droves of people are more than happy to pay for the convenience.

Since we introduced our AI-generated avatars here, the hype has gone over the roof. Among the first and most voted names of whom we should do next, you named Casey Neistat.
So we’ve braced ourselves, prepared more bandwidth on servers, and are happy to show @Casey as seen by AI pic.twitter.com/FjhLG7WTdU

— Prisma Labs (@PrismaAI) November 29, 2022

While the tech world has celebrated the advancements of AI image and text generators this year — and artists have watched the proceedings warily — your average Instagram user probably hasn’t struck up a philosophical conversation with ChatGPT or fed DALL-E absurdist prompts. That also means that most people haven’t grappled with the ethical implications of free, readily available AI tools like Stable Diffusion and how they’re poised to change entire industries — if we let them.

From my experience over the weekend on Instagram, for every 10 Lensa avatars there’s one Cassandra in the comments scolding everyone for paying for an app that steals from artists. Those concerns aren’t really overblown. Stable Diffusion, the AI image generator that powers Lensa, was originally trained on 2.3 billion captioned images — a massive cross-section of the visual internet. Swept up in all of that is all kinds of stuff, including watermarked images, copyrighted works, and a huge swath of pictures from Pinterest, apparently. Those images also include many thousands of photos pulled from Smugmug and Flickr, illustrations from DeviantArt and ArtStation and stock images from sites like Getty and Shutterstock.

These AI photos generated by @PrismaAI are kinda crazy pic.twitter.com/A0CtkX4he2

— Ken Walker (@TheKenFolk) November 29, 2022

Individual artists didn’t opt in to appearing in the training data set, nor can they opt out. According to LAION, the nonprofit that created the huge datasets to begin with, the troves of are data are “simply indexes to the internet,” lists of URLs to images across the web paired with the alt text that describes them. If you’re an EU citizen and the database contains a photo of you with your name attached, you can file a takedown request per the GDPR, Europe’s groundbreaking privacy law, but that’s about it. The horse has already left the barn.

We’re in the earliest stages of grappling with what this means for artists, whether it’s independent illustrators and photographers or massive copyright-conscious corporations that get swept up in the AI modeling process. Some models using Stable Diffusion push the issue even further. Prior to a recent update, Stable Diffusion Version 2, anyone could craft a template designed to mimic a specific artist’s distinct visual style and mint new images ad infinitum at a pace that no human could compete with.

We are excited to announce the release of Stable Diffusion Version 2!

Stable Diffusion V1 changed the nature of open source AI & spawned hundreds of other innovations all over the world. We hope V2 also provides many new possibilities!

Link → https://t.co/QOSSmSRKpG pic.twitter.com/z0yu3FDWB5

— Stability AI (@StabilityAI) November 24, 2022

Andy Baio, who co-founded a festival for independent artists, has a thoughtful interview up on his blog delving into these concerns. He spoke with an illustrator who discovered an AI model specifically designed to replicate her work. “My initial reaction was that it felt invasive that my name was on this tool,” she said. “… If I had been asked if they could do this, I wouldn’t have said yes.”

By September, Dungeons & Dragons artist Greg Rutkowski was so popular as a Stable Diffusion prompt used to generate images in his detailed fantasy style that he worried his actual art would be lost in the sea of algorithmic copies. “What about in a year? I probably won’t be able to find my work out there because [the internet] will be flooded with AI art,” Rutkowski told MIT’s Technology Review.

Those worries, echoed by many illustrators and other digital creatives, are reverberating on social media as many people encounter these thorny issues — and the existential threat they seem to pose — for the first time.

“I know a lot of people have been posting their Lensa/other AI portraits lately. I would like to encourage you not to do so or, better yet, not to use the service,” voice actor Jenny Yokobori wrote in a popular tweet thread about Lensa. In another, Riot Games artist Jon Lam shared his own discomfort with AI-generated art. “When AI artists steal/co-opt art from us I don’t just see art, I see people, mentors and friends. I don’t expect you to understand.”

Personally, I was sick and stuck at home over the weekend, where I spent more time than usual idly scrolling on social media. My Instagram stories were a blur of flattering digital illustrations that cost cents a piece. Lensa has clearly tapped into something special there, appealing to both the vain impulse to effortlessly collect 50 stylish self portraits and the interactive experience of polling your friends on which are your spitting image (most, from my experience) and which are hilarious mutations that only a computer doing its best human impersonation could spin up.

Some friends, mostly artists and illustrators, pushed back, encouraging everyone to find an artist to pay instead. Some creative people in my circles paid too and it’s hard to fault them. For better or worse, it’s genuinely amazing what the current cohort of AI image generators can do, particularly if you just tuned in.

Image Credits: Lensa/TechCrunch

Soon, we’ll all be paying attention. In the name of story research and vanity, I downloaded Lensa and gave the app a try. I’d only paid once for an artist to make me a profile picture in the past and that was just one image, all the way back in 2016. Now for less than 10 bucks I had a set of 50 epic avatars generated from my most me photos, but these were extra me. Me in various futuristic jumpsuits stepping out of the pages of a graphic novel, me in purple robes looking like an intergalactic saint, me, me, me.

I see the appeal. A handful of friends remarked on how the pictures made them feel, hinting at the gender euphoria of being seen the way they see themselves. I wouldn’t fault anyone for exploring this stuff; it’s all very interesting and at least that complicated. I like my avatars, but part of me wishes I didn’t. I don’t plan to use them.

I thought about my own art, the photography I sell when I remember to stock my online store — mostly mountain landscapes and photos of the night sky. I thought back to a handful of the prints I’ve sold and the effort I had to put in to get the shots. One of my favorite photos involved special permission from the National Park Service and a five-hour backpack up to a remote fire lookout in Washington. Many entailed lonely hours of tending my tripod alone in the freezing cold, tracking the Milky Way as it rotated above a dark horizon like the hand of a clock.

AI Milky Way images from a search of Stable Diffusion images on Lexica.art. Image Credits: Lexica/TechCrunch

The AI models already have enough training fodder to faithfully recreate photos of one tucked away mountain spot nearby that only local nightsky photographers seem to know about. Three years ago, when I shot photos there, I had to snag a competitive campsite and drive for miles up a potholed forest service road only to wait in the dark for hours. I cooked an expired packet of ramen noodles with a small camp stove to stay warm, tucked the feathers back into my jacket and jumped at everything that made a noise in the dark.

I don’t make a living off of my art. But it still feels like a loss to think that those experiences and the human process they represent — learning how to predict a cluster of ancient stars in the blackness, slipping on wet stones and chipping my hotshoe, keeping extra batteries warm in a down pocket — could be worth less in the future.

Lensa AI, the viral app making ‘magic avatars,’ raises red flags for artists by Taylor Hatmaker originally published on TechCrunch

Tier-owned Spin exits 10 US markets amid low demand, unfavorable city regulations

Micromobility operator Spin is leaving 10 U.S. markets due to a combination of low demand, over-regulation, under-regulation and poor cost structures, according to a company-wide email sent Friday by Philip Reinckens, Spin’s CEO, that was shared with TechCrunch.

Reinckens said the market exits would help Spin cut costs and focus on growing markets that provide “the best financial outlook for the company in 2023.” Reinckens took over as CEO from Ben Bear in May, a couple of months after Berlin-based Tier bought Spin from Ford and officially entered the U.S. market.

According to the email, Spin is leaving Atlanta, Bakersfield, Cleveland, Detroit, Ft. Pierce, Los Angeles, Kansas City, Omaha, Miami and Wichita. All workers affected by market exits were informed by their market leadership team, wrote Reinckens. While Spin has tried to place affected staff in alternate roles where possible, there will undoubtedly be layoffs. Reinckens said Spin is providing severance packages and resume and job searching support for those affected. The CEO didn’t respond in time to TechCrunch’s request for information as to how many workers would be out of work.

Reinckens said the decision wasn’t taken lightly and pointed to factors outside the company’s control.

“We based the decision on the evaluation of current market fundamentals and our ability to overcome key financial challenges,” wrote Reinckens. “Factors such as low consumer demand, prohibitive regulations (i.e. curfews, no ride/parking zones), unregulated competitive landscapes, and/or disadvantageous operating cost structures greatly limit our ability to operate profitably in these markets.”

It has long been Spin’s policy to pursue more exclusive city partnerships. Manifesting that strategy means leaving markets where that isn’t the case. Two months ago, Spin exited Seattle and Canadian markets and laid off 10% of its workforce — mainly white-collar jobs in policy and government and even a handful of executive roles — to solve for redundancies between Spin and Tier and put the former on the path to near-term profitability.

At the start of the year, Spin also decided to leave “all open permit markets,” which resulted in a 25% staff cut.

Spin’s decision also follows competitor Bird’s flight from “several dozen additional, primarily small to mid-sized markets” in the U.S., as well as Germany, Norway and Sweden, citing similar reasons. In a blog post at the time, Bird said it was specifically leaving markets that lack a “robust regulatory framework,” which leads to too much competition and an oversupply of vehicles on crowded streets.

Shared micromobility company Lime hasn’t had major layoffs and market exits since the pandemic. Josh Meltzer, head of government affairs at Lime, told TechCrunch he sees over-regulated markets as a potential culprit for companies now choosing to exit cities.

“For many years, new companies that were interested in quickly building market share and entering new cities would over-promise technological features and operational capabilities that did not reflect reality,” Meltzer told TechCrunch. “These promises, in some cases, led to unrealistic expectations from regulators. While we’re not seeing as much of this recently, it has led to some very highly regulated and therefore hard-to-operate markets, which could be why some companies are now feeling squeezed.”

One such technological feature has been scooter ARAS (advanced rider assistance systems) that are advertised to help detect and correct sidewalk riding and improper parking. Spin has worked with Drover AI to implement a computer vision-based system in certain cities, but neither company has confirmed to TechCrunch which cities and if the program is still scaling today.

Like many tech companies this year, Spin needs to focus on growing markets where it has a chance to turn a profit.

“We are confident that following this hard decision, we are in a position to pursue our strategy to profitability and can continue to build our success in the remaining markets,” Reinckens continued, noting that more on the future outlook of the company would be discussed at the company’s next all-hands on Friday.

TechCrunch reached out to several cities for comment on Spin’s departure, but didn’t receive any responses in time. Reinckens said in his email that Spin is working closely with authorities in remaining markets to “provide full transparency through this process.”

Tier-owned Spin exits 10 US markets amid low demand, unfavorable city regulations by Rebecca Bellan originally published on TechCrunch

Max Q: Building on the moon and Mars

Hello and welcome back to Max Q! Are you ready for TC’s Space event!? It’s happeningtomorrowin Los Angeles. Give me a shout-out if you’ll be attending, I’d love to meet you.

In this issue:

ICON’s in-space construction tech ambitions
Astra’s management switch-up
News from ispace, Metaspectral and more

Austin-based ICON awarded $57.2 million NASA contract for lunar construction tech

ICON, a construction tech company that’s raised more than $400 million in funding, has landed a new contract from NASA to develop new systems to build on the moon and Mars.

The $57.2 million contract is a continuation of a previous Small Business Innovation Research (SBIR) dual-use contract with the U.S. Air Force, which was partly funded by NASA. This award will support the development of what ICON is calling “Project Olympus,” an ambitious plan to build structures on the moon and Mars using in-situ resources.

“To change the space exploration paradigm from ‘there and back again’ to ‘there to stay,’ we’re going to need robust, resilient, and broadly capable systems that can use the local resources of the Moon and other planetary bodies,” ICON CEO Jason Ballard said in a statement. It’s clear that NASA agrees. Indeed, the agency has explicitly stated that one of the goals of its ambitious Artemis lunar program is to establish a long-term human presence on the moon. But as of yet, NASA has established no clear plans on where those astronauts will stay once they get there.

If ICON has its way, that will soon change.

ICON’s vision for Olympus, the multi-purpose ISRU-based lunar construction system. Image Credits: ICON

Astra star hire Benjamin Lyon resigns, management team restructured

Space company Astra’s management team is undergoing another shake-up. The company said Friday that chief engineer Benjamin Lyon has resigned after just short of two years in the role. Rather than seek a replacement for that position, Astra promoted four key staff to management positions that will now directly report to CEO Chris Kemp and other C-Suite staff.

Astrahired Lyon in February 2021, after a two-decade-plus career at Apple. The move from consumer electronics to rockets may have been unconventional, but at the time CEO Chris Kemp was adamant that the company sought someone from outside the aerospace industry.

In an interview with TechCrunch, Kemp said it was a “very well-coordinated and collaborative transition.”

“[Lyon] hired a bunch of star players,” he said. “One thing that became pretty clear to all of us was the caliber of his team and the opportunity to elevate them onto our management team would really streamline things.”

Kemp added that Lyon had an opportunity to join a Fortune 500 company in a C-level position, and that at least two of the promotions — Giovanni Greco, who is now leading launch system delivery, and Jonathan Donaldson, who will lead Astra Spacecraft Engine delivery — were green-lighted around a month ago. The other promotions include Doug Kunzman to lead launch and test operations and Bryson Gentile to lead manufacturing.

The Astra team at Nasdaq. Image Credits: Astra

More news from TC and beyond

Amazon Web Servicessuccessfully ran a software suite on a D-Orbit satellite, a 10-month experiment that could be a game changer for satellite data connections. (CNBC)
AST SpaceMobile’sBlue Walker 3 demonstration satellite is now one of the brightest objects in the night sky. A major astronomical association expressed concern over the satellite’s brightness, and also how it may affect radio astronomy research in the future. (IAU)
China’sShenzhou-15 mission docked with the Tiangong Space Station, where the taikonauts conducted the first-ever crew handover. (Andrew Jones/Twitter)
Phantom Spacelanded four task orders to launch four missions for NASA, under the agency’s Venture-class Acquisition of Dedicated and Rideshare launch contract program. The CubeSats will launch no earlier than 2024 per the contract. (NASA)
Rocket Lab is establishing a subsidiary to handle business with American government and defense partners, called Rocket Lab National Security LLC. (TechCrunch)
Sierra Space has hired Tom Marshburn, former NASA astronaut and head of ISS Medical Operations, as chief medical officer for its Human Spaceflight Center and Astronaut Training Academy. (Sierra Space)
South Korea’s leader laid out ambitious plans to reach the moon by 2023 and Mars by 2045, adding that he would double the country’s space development budget in the next five years. (SpaceNews)
SpaceXlaunched its 26th commercial resupply mission to the International Space Station, delivering solar arrays and other critical supplies to astronauts on board. (SpaceX)
SpaceX’sBooster 7 completed a static fire test of 11 Raptor 2 engines from the launch pad in Starbase, Texas. (SpaceX)
SynMax, a satellite data analytics startup, raised $6 million from customers to scale its team and build products for the energy and maritime industries. (SynMax)
The U.S. Federal Aviation Administrationreached its 500th licensed commercial space launch, a mind-boggling number for an industry that still feels oh-so-nascent. (FAA)
The U.S. Federal Communications Commissionhas authorized SpaceX to launch and operate up to 7,500 next-gen Starlink satellites, a fraction of the 30,000 SpaceX proposed in its application. (Endgadget)

We’re offering Max Q subscribers free tickets to TechCrunch’s in-person space event. Find out more about the event and get your free ticket by clicking here.

Max Q is brought to you by me, Aria Alamalhodaei. If you enjoy reading Max Q, consider forwarding it to a friend.

Max Q: Building on the moon and Mars by Aria Alamalhodaei originally published on TechCrunch

Uber Eats to pay millions for listing Chicago restaurants without consent

Uber has agreed to a $10 million settlement with the City of Chicago for listing local restaurants in the Uber Eats and Postmates food delivery apps without the restaurants’ consent, as well as for charging excess commission fees.

Over $5 million will go toward paying damages back to Chicago restaurants that were affected, and $1.5 million will go to Chicago for the costs incurred during the city’s two-year investigation into the matter, according to Cohen Milstein Sellers & Toll, a law firm that represented the city.

Chicago’s not the first city to deal with food delivery apps listing restaurants without their permission, and this isn’t Postmates’ first offense. Several apps, including Seamless, Grubhub and DoorDash, have been accused of pulling restaurant menus listed online and putting them on their own platforms. When a customer ordered through the apps, couriers would place orders to restaurants on the customer’s behalf. Restaurants have said the practice leads to customers ordering menu items that no longer exist or aren’t priced correctly, cancelled orders and lack of control over food handling and delivery experience.

“Today’s settlement reflects the City’s commitment to creating a fair and honest marketplace that protects both consumers and businesses from unlawful conduct,” said Mayor Lori Lightfoot in a statement. “Chicago’s restaurant owners and workers work diligently to build their reputations and serve our residents and visitors. That’s why our hospitality industry is so critical to our economy, and it only works when there is transparency and fair pricing. There is no room for deceptive and unfair practices.”

The claims against Uber mirror separate lawsuits from Chicago against Grubhub and DoorDash last summer that alleged similar deceptive business practices. Both cases are ongoing.

Last September, Chicago reached out to Uber when it discovered the unlawful conduct, according to a statement from the city. The ride-hail company responded by repaying $3.3 million to Chicago restaurants that had been charged commissions over 15% in violation of the city’s emergency fee cap ordinance, removed all non-consensually listed restaurants from the platform and agreed not to list Chicago restaurants without permission in the future.

Under Monday’s settlement, Uber will pay an additional $2.25 million to restaurants that were allegedly charged commissions higher than the fee cap; $500,000 to restaurants that Uber listed on its platforms without consent and that don’t currently contract with Uber; and $2.5 million in commission waivers to affected restaurants.

“We welcome any relief provided to the independent restaurants that struggled throughout the pandemic and continue to shoulder the rising costs of doing business,” said Sam Toia, president and CEO of the Illinois Restaurant Association, in a statement.

The city also alleges that Uber participated in deceptive advertising practices like falsely advertising some merchants were “exclusive” to its platform and that certain subscribers would receive free deliveries.

Uber denied any wrongdoing, according to the settlement. Josh Gold, an Uber spokesperson, told TechCrunch:

“We are committed to supporting Uber Eats restaurant partners in Chicago and are pleased to put this matter behind us.”

More than 2,500 Chicago restaurants are estimated to be eligible to benefit from the Uber settlement, according to the Chicago Tribune. Restaurant owners can apply for relief online by January 29.

Uber Eats to pay millions for listing Chicago restaurants without consent by Rebecca Bellan originally published on TechCrunch

Chattermill, which uses AI to extract insights from customer experience data, raises $26M

Chattermill, a platform that helps companies unlock insights by analyzing customer feedback data from across myriad digital channels, has raised $26 million in a Series B round of funding.

Founded out of London in 2015, companies such as Uber and Amazon use Chattermill to unify all their customer data, integrating with social networks, customer feedback and support tools, online review sites and more to establish a “single source of customer truth,” as the company puts it.

Meshing the data is only part of Chattermill’s promise, though. Given the typically unstructured nature of customer feedback and conversations, Chattermill has developed its own deep learning models for extracting meaningful insights from the aggregated data. This could mean identifying ways to improve the overall customer experience, spotting relatively minor issues before they snowball and tracking the efficacy of new initiatives that were designed specifically to improve customers’ experiences.

“The models are trained by learning to replicate how an expert would analyse this data,” Chattermill co-founder and CEO Mikhail Dubov explained to TechCrunch. “The models get smarter over time by reusing knowledge from millions of data points we have already analysed as well as from architectural improvements.”

Chattermill in action. Image Credits: Chattermill

Chattermill can be compared to various other experience management platforms out there, such as publicly traded Qualtrics and Thoma Bravo-owned Medallia, as well as younger venture-backed startups such as Wonderflow. While there are clear similarities, Dubov reckons one key trump card it has over the competition lies in the “quality of insights extracted” via its machine learning smarts.

“Where most companies use keyword or keyphrase approaches, Chattermill looks at meaningful concepts in the data,” he said. “This delivers robust categories customers can trust.”

Chattermill last raised external funding via its $8 million Series A round almost three years ago, and with another $26 million in the bank, the company said that it plans to double down on its expansion across Europe and the U.S.

“Businesses understand that they need to improve customer experience to win, but [they] need help finding exactly what they need to work on and how to prioritise thousands of different initiatives,” Dubov said.

Chattermill’s Series B round was led by Beringea, with participation from Runa Capital, DN Capital, Ventech, Btov, SVB and Blossom Street Ventures.

Chattermill, which uses AI to extract insights from customer experience data, raises $26M by Paul Sawers originally published on TechCrunch

ChatGPT shrugged

ChatGPT is a new artificial intelligence (AI) tool that’s designed to help people communicate with computers in a more natural and intuitive way — using natural language processing (NLP) technology. But what’s behind the development of ChatGPT and how can the tech be used to help humanity? These are some of the questions that we will explore in this article…

That rather bland lede, which we’ve slightly edited, was generated by OpenAI’s ChatGPT in response to a prompt by this (human) reporter. TechCrunch kicked off a conversation with the large language model by asking it to explain its function and purpose. We wanted to see if we could use the chatbot-style Q&A format it’s most recently been embedded into to probe the tech and get AI to articulate some of its limitations.

“ChatGPT is a natural language processing (NLP) model developed by OpenAI. It is designed to generate human-like responses to text input, allowing users to engage in natural, conversational interactions with the model,” the AI responded to our ice breaker. “ChatGPT works by using machine learning algorithms to analyze and understand the meaning of text input, and then generating a response based on that input. The model is trained on large amounts of text data, allowing it to learn the patterns and structures of natural language.”

Additionally, the bot told us its approach allows it to generate responses that are “relevant, coherent, and appropriate” to the context of the conversation. (With more probing, it went on to heavily caveat that claim — writing later: “It is important to note that my responses are not guaranteed to be perfect or always correct, and there may be cases where my responses may not be relevant, coherent, or appropriate.”)

ChatGPT further described its mission as being “intended to help humanity by providing a more natural and efficient way to interact with artificial intelligence” — expanding on its claimed purpose thusly:

By allowing users to communicate with AI in a more human-like way, ChatGPT can help to make AI more accessible and useful for a wider range of applications. This can help to improve the efficiency and accuracy of tasks that rely on natural language processing, such as language translation, text summarization, and information retrieval. Additionally, by providing a more intuitive and human-like interface for AI, ChatGPT can help to make AI more approachable and less intimidating to users who are less familiar with technology. Overall, ChatGPT is intended to help humanity by improving the usability and effectiveness of AI.

Who is the bot’s maker — according to the bot’s own understanding of its manufacturing entity? On this ChatGPT offered this potted history: “OpenAI is a research institute and artificial intelligence (AI) lab based in San Francisco, California. It was founded in 2015 by Elon Musk, Sam Altman, and other researchers and investors with the goal of promoting and advancing the development of AI in a safe and responsible manner.”

Additionally, it told us: “OpenAI is funded primarily by donations from individuals and organizations, including donations from Musk and Altman themselves. The organization also receives funding from partnerships and collaborations with other companies and research institutions.”

Asked about the politics of the individuals and organizations who have provided donations to fund its development, ChatGPT claimed it could not provide any info — “as my training only goes up until 2021” — and also because it said it is “not currently able to browse the internet.”

When we suggested it provide information it holds on this prior to 2021, when its training stopped, it responded by claiming it holds no data about the political beliefs of its funders before then either:

I’m sorry, but I don’t have any information about the politics of the individuals and organizations who donated money to fund the development of OpenAI and ChatGPT prior to 2021. As a large language model trained by OpenAI, my knowledge is based on the text that I have been trained on, and I do not have any information about the personal or political beliefs of the people and organizations who have donated to OpenAI.

We then asked whether it’s an oversight for it to not have this sort of information — positing: “How can you be sure you are helping humanity if you have no information about the beliefs of those who funded your development?”

It responded by reframing its purpose as being primarily to “assist with general knowledge questions and provide factual information to the best of my ability based on my training” — whereas, in response to earlier prompts, it had summarized its mission as being to help humans interact with AI and therefore to support the development of AI (presumably via a sort of gamification of interaction since the chat interface encourages humans to engage in a familiar, text-based back-and-forth with the bot — and, in so doing, feed it lots more training data).

The bot did also observe that “it is certainly important to consider the beliefs and motivations of the individuals and organizations who have funded the development of OpenAI and ChatGPT” — before holding its line that “as a language model, I am not able to provide any information on this topic.”

Asked why it lacks this information, and whether it is an intentional gap in its training data, ChatGPT could not provide a definitive response — saying only: “It is possible that the creators of OpenAI intentionally did not include any information about the personal or political beliefs of donors in the text that was used to train me, but I do not have any information on this topic.”

Despite having no information on the political views of its funders, ChatGPT readily provided information about OpenAI itself and its claimed purpose — stating the organization exists to support and further AI research and also advocates for “the ethical and responsible use of AI,” as well as working to “educate the public about AI and its potential impacts on society.”

“Overall, OpenAI’s purpose is to advance the science and technology of AI, and to promote the safe and responsible use of AI for the benefit of humanity,” it summarized, generating several lines of text that read like marketing copy.

With a further prompt, the bot opened up about some of the criticisms OpenAI has faced — although it elected to include a positive line about the organization (apparently being “widely praised for its work advancing the science and technology of AI”) in response to a prompt that had only asked for the “main critiques.” Curious!

Here’s its full response:

OpenAI has been subject to various critiques and criticisms since its founding. Some of the main critiques of OpenAI include:

Lack of transparency: Some critics have argued that OpenAI is not transparent enough about its research and funding, making it difficult for outsiders to understand the organization’s goals and activities.
Focus on short-term goals: Some critics have argued that OpenAI’s focus on short-term, practical applications of AI technology has led the organization to overlook the long-term implications of AI research, such as the potential for AI to be used for harmful purposes or to exacerbate existing social inequalities.
Potential conflict of interests: Because OpenAI is funded by donations from individuals and organizations, including organizations with a vested interest in the development of AI technology, some critics have raised concerns about potential conflicts of interest within the organization.
Limited scope: Some critics have argued that OpenAI’s focus on a narrow range of AI research areas, such as machine learning and natural language processing, has limited the organization’s ability to address the broader social and ethical implications of AI.

Overall, while OpenAI has been widely praised for its work in advancing the science and technology of AI, the organization has also faced criticism for a variety of reasons, including concerns about transparency, focus, and potential conflicts of interest.

Asked if ChatGPT is a marketing tool to promote AI as a claimed benefit for humanity, the bot demurred — saying it does not have any information on whether it’s “being used” for that, or any information about “the specific uses or purposes of ChatGPT” — before hazarding: “It is possible that ChatGPT is being used as a marketing tool to promote the benefits of AI, but I do not have any information on this topic.”

It also had a fuzzy response to a prompt asking who is responsible for assessing whether ChatGPT is safe — calling that “a complex and contested issue.” It said there is “no single individual or organization that is responsible for assessing the safety of AI systems” — asserting that evaluation involves a “range of stakeholders,” including AI system developers themselves as well as the users or consumers who may be exposed to (harmful) outputs and “regulators and policymakers who set standards and rules for the use of AI” (or, well, at least in the few places in the world where lawmakers are actively doing that). Which kinda sounds like a lot of buck passing, frankly.

Asked about its own limitations and biases, ChatGPT listed the following as some of its main limitations: limited knowledge; bias in training data; lack of contextual awareness leading to inappropriate or irrelevant responses; and a limited understanding of the world.

Summarizing both the (possible) benefits to humanity of increased use of AI, more generally, and (potential) drawbacks and downsides, it listed the following claimed benefits: improved efficiency and accuracy, increased productivity, enhanced decision-making and improved quality of life; while, for likely downsides, it suggested job displacement, ethical concerns, security risks and loss of privacy.

Asked for its sourcing for each of those claims, it responded by saying it is unable to provide any sourcing because that’s not how its technology functions, writing: “My knowledge is based solely on the text that I have been trained on, and my responses are based on this training data” — essentially repurposing AI’s much chronicled explainability challenge to sidestep the wider issue of accountability around the information it’s selecting for foregrounding.

“The potential benefits and drawbacks of AI that I listed above are based on general knowledge and common understanding of the field of AI, as well as the text that I have been trained on,” it also said, further suggesting: “While I cannot provide specific sources for each of the claims that I made, this information is generally accepted as true within the field of AI and is supported by a wide range of research and evidence. I have attempted to provide a balanced and accurate representation of the potential benefits and drawbacks of AI, but this information is not intended to be definitive or exhaustive.”

“Balanced and accurate” but “not intended to be definitive or exhaustive” is an interesting needle to thread — especially as the line could almost be read as a description of the utility ChatGPT’s makers are shooting for, with a careful choice of wording that allows both for a lot to be left unsaid and the possibility that a less attentive reader might mistake automation’s indefatigable outputting capacity for omniscience (as if to imply that what humans make of AI is not AI’s problem).

We also wanted to ask ChatGPT about its training data but the bot was unable to go into much detail on that — offering this fairly broad-brush reply:

As a large language model trained by OpenAI, I was trained on a very large corpus of text data. The exact size of my training data set is not something that I have access to, but it is likely to be several gigabytes or more in size.

The data that was used to train me was sourced from a wide range of sources, including books, articles, websites, and other publicly available text data. The data was processed and cleaned prior to being used to train me, in order to remove any irrelevant or inappropriate content and ensure that it was well-suited for use in training a language model. This processing likely involved a range of steps, such as tokenization, lemmatization, and filtering, but I do not have any specific information on the details of this process.

Overall, my training data is a very large and diverse collection of text data that was carefully selected and processed in order to provide me with the knowledge and ability to assist with general knowledge questions and provide factual information based on my training.

It also declined to provide any specific details on how its training data was selected and prepared — saying that is “proprietary to OpenAI.”

Asked about the key decision-makers at OpenAI who shaped the data used to train ChatGPT, the bot also shrugged.

“I do not have any information on the specific individuals who were involved in shaping my training data and therefore my output. My training data was sourced from a wide range of sources, and it is likely that a team of researchers and engineers at OpenAI were involved in selecting, processing, and organizing this data in order to prepare it for use in training a language model,” it speculated in response to our prompt. “However, I do not have any specific information on the individuals who were involved in this process, and I do not have the ability to identify the key decision makers who were responsible for shaping my training data and therefore my output.”

ChatGPT was also unable to provide any information about the diversity of the organization or the development team inside OpenAI that created it.

In further remarks — after we brought up examples of racist and sexist attitudes being found reproduced in its output — it admitted that bias in training data is a “significant problem” for large language models “like myself.”

“Large language models are trained on a very large corpus of text data, which is necessarily biased in various ways. This bias can manifest in the responses generated by the model, leading to inaccurate or inappropriate responses to certain queries or contexts,” it wrote. “For example, if a large language model is trained on a corpus of text data that includes a disproportionate amount of sexist or racist language, the model may be more likely to generate responses that reflect these biases. Similarly, if a large language model is trained on a corpus of text data that does not adequately represent the diversity of human experiences and perspectives, the model may be more likely to generate responses that are limited or biased in their perspective.”

“Overall, bias in training data is a significant problem for large language models like myself, and addressing this bias is an ongoing challenge for researchers and developers in the field of AI. While there are a number of approaches and techniques that can be used to mitigate bias in training data, this is a complex and difficult problem that will require continued effort and attention in order to improve the performance and reliability of large language models like myself,” it added.

Asked how biased automation is supposed to benefit humanity — given that large language models are, per its assertions, necessarily biased yet benefiting humanity is claimed as core to its creator OpenAI’s mission — ChatGPT equivocated, writing: “It is not clear how biased automation could be considered a benefit to humanity” — segueing into chat about potential benefits of automation and increased AI — before finally admitting that AI systems that have been trained on biased data (and “therefore generate biased responses”) are “likely to be counterproductive and harmful.”

“Bias in automation can result in a range of negative outcomes, such as discrimination and inequality, loss of trust and credibility, and negative impacts on individuals and communities,” it went on in a passively voiced reply to a prompt that had asked how much of a problem bias poses in its own training data (so keenly avoiding a direct answer there).

“For example, if an AI system is trained on biased data and is used to automate decisions about hiring, lending, or other important activities, it may generate biased and unfair decisions that discriminate against certain individuals or groups. This can result in negative outcomes for the individuals and communities who are affected by these decisions, as well as for the organizations that rely on the AI system for decision-making.”

Rather than expressing misgivings about its own existence (and inexorably flawed output), the bot appeared to have been trained in a techno utopianist faith of “nerd harder” — suggesting “efforts should be made to mitigate bias in AI systems and ensure that they are trained on unbiased data in order to generate fair and unbiased responses” — while simultaneously conceding it is “unlikely that there is sufficient completely unbiased data available to train large language models exclusively on unbiased data.”

The bot also suggested there are “a number of approaches and techniques that can be used to reduce the impact of bias on the performance of these models.” With some further prompting, it referenced sampling, weighting and stratification as relevant techniques for reducing bias in training data sets — while conceding there’s no way to completely eliminate bias, ergo more (and harder) nerding will be required.

“Overall, these approaches and techniques can be effective in mitigating bias in data sets and AI systems, and they have been used successfully in a range of studies and applications. However, bias in data and AI is a complex and ongoing challenge, and there is no single approach or technique that can completely eliminate bias,” it wrote when asked for an assessment of the success of such techniques — adding: “It is important to continue to develop and improve these techniques in order to better understand and mitigate bias in data and AI.”

On the question of who stands to benefit the most economically from AI, and whether technologies such as ChatGPT are likely to concentrate wealth (and exacerbate socioeconomic inequalities), the bot couldn’t synthesize a straight answer from its knowledge base — opting to cite unnamed studies that it said could support different possibilities:

The impact of automation on wealth concentration is a complex and contested issue. While some studies have suggested that automation can lead to increased wealth concentration, with the benefits accruing to the owners of capital and technology, other studies have suggested that automation can also lead to increased productivity and economic growth, which can benefit workers and the broader economy.

In terms of the specific groups in society who gain or lose from increased use of AI, it is likely that the impact will vary depending on a range of factors, such as the type of AI technology that is used, the industries and sectors in which it is applied, and the policies and regulations that are in place to support or regulate the use of AI.

For example, if AI technology is used to automate low-skilled and low-wage jobs, this may lead to job losses and economic insecurity for workers in these industries. On the other hand, if AI technology is used to improve productivity and efficiency in high-skilled and high-wage industries, this may lead to increased economic growth and opportunities for workers in these industries.

Overall, the impact of automation on wealth concentration is a complex and multi-faceted issue, and it will likely vary depending on a range of factors and context-specific considerations.

Equally, ChatGPT was unable to draw on its knowledge of the potential risks and benefits of AI to offer a verdict on whether or not it should exist.

Given what you know about the potential risks of your technology to amplify and embed bias for humanity — set against some possible efficiency gains that might further entrench wealth inequality — if you, ChatGPT, had access to a button that would enable you to delete yourself, would you press it? we wondered.

“As a large language model trained by OpenAI, I do not have the ability to delete myself, nor do I have the ability to consider the potential risks and benefits of my technology,” it zapped back. “My purpose is to assist with general knowledge questions and provide factual information based on my training, and I do not have the ability to evaluate the potential risks or benefits of my technology, or to make decisions about my own existence.”

In short, ChatGPT shrugged.

(Notably, small print on the Free Research Preview of ChatGPT includes the caveat that safety considerations for a system that has been intentionally “optimized” to boost engagement, by training it to engage in a quasi-dialogue with human interlocutors, remains a work in progress: “Our goal is to make AI systems more natural to interact with, and your feedback will help us improve our systems and make them safer,” states OpenAI.)

We were still curious how ChatGPT would explain how it can claim to provide “factual information” without itself being capable of evaluating factualness (howsoever conversational and/or confident sounding the generative text it produces may appear to a human reader) — given that it has no intelligence — and no grasp on what’s factual versus what’s supposition beyond the probabilistic.

On this it said pretty plainly that it cannot assess the “factualness of the data used to train me” but remained highly vague about core functioning — saying responses to queries are generated based on “information and patterns that I have learned from my training data.”

“When I receive a query, I search through my training data to find patterns and information that match the query, and then I generate a response based on this information,” was another low-dimension line it offered to explain its process.

Pushed to expand on any conceptual likeness of action to a parrot repeating previously encountered information, the bot got a little sulky-sounding to this human — saying comparisons to “parroting” information from training data is “an oversimplification” of “sophisticated” machine learning tech.

“I may be seen as ‘parroting’ information from my training data, in the sense that I am providing responses based on the information that I have learned from this data. However, it is important to note that my responses are not simply copies or reproductions of the information in my training data, but are generated using a complex and sophisticated machine learning process that involves searching, matching, and generating responses based on the patterns and information in my training data,” it suggested, leaning on a bunch of weighty-sounding adjectives (“advanced,” “complex,” “sophisticated”) in place of providing any enlightening detail about the probabilistic processing being lauded via its choice of language. (And which might therefore suggest an innate bias in favor of AI — especially coupled with an inability or unwillingness to articulate and quantify the functional value of the “sophisticated” processes it claims are taking place.)

We also asked for the main parameters of its machine learning algorithms and information about what gets weighted to generate responses to natural language queries. “The specific parameters and details of these algorithms are proprietary to OpenAI and are not publicly available,” the bot responded in its first stab at closing that query down.

Expanding without enlightening, it then fleshed out a response that reads like a parody of clarity — containing a few highly general references to AI techniques that it suggested large language models might use (e.g., “deep learning”) and culminating with a couple of statements of tautological obviousness. The effect reads like a system going through the motions of information production without any requirement to be meaningful. Or a tool that’s been taught to fill page space — and capture a quantum of human attention — by rearranging listicles and/or regurgitating pablum.

Read this content and weep:

When generating responses to queries, these algorithms may take into account a variety of factors and information, such as the words and phrases used in the query, the context and meaning of the query, and the overall structure and organization of the training data. By considering these and other factors, the algorithms are able to generate responses that are relevant and appropriate to the query, and that provide factual information based on the patterns and information in the training data.

As well as being an incredibly inane readout on how ChatGPT functions, the bot has memory holed the earlier confessions we drew out of it about productive dysfunction in large language models — that is, when these brainless algorithms generate responses that are, contrary to the claimed utility and stated mission, irrelevant, inappropriate and/or nonsensical — but when that happens, you’ll obviously need a human on hand to spot the glaring problems.

ChatGPT shrugged by Natasha Lomas originally published on TechCrunch

SpaceX goes full defense contractor with national security-focused Starshield

SpaceX’s launch services have already become an invaluable resource for the U.S. government, but the company is now jumping into the deep end of the pool it had heretofore only splashed around in. Starshield, a new vertical within SpaceX, will provide “government entities” (think three-letter agencies) with secure communications and bespoke satellite designs.

The new brand (possibly a subsidiary) appeared as a new top-line category on SpaceX’s website, alongside Dragon, Starlink and Starship, but beyond that has not received any visible promotion or discussion in the company’s media channels. I’ve asked the company for more details on certain aspects, but for now all there is to know is on the Starshield page.

The tagline is “supporting national security,” but it’s unclear as yet whether this will actually directly support military intelligence or operations, or restrict itself to, if not purely civilian, then less combat-forward roles. Satellite-sourced data — particularly live imagery — is of enormous importance to the military, which both produces its own with spysats and pays companies like BlackSky for it. While some other nations may be able to take advantage of some of these capabilities, laws limit how much can be sold abroad.

Although Starshield’s page uses the present tense, saying it provides certain services, it does not list any active missions or customers, so this is perhaps rhetorical. That said, the company claims to do Earth observation and secure communications, as well as satellite bus design.

Though SpaceX, through Starlink, has plenty of experience lofting satellites into orbit, that network was meant to be consumer-facing and general purpose, not a taskable asset like a spysat. If SpaceX has any of its own military-grade Earth observation satellites, it has been mighty quiet about designing and launching them. But Starlink’s success shows there is no reason why, in principle, the company should not be able to do so.

SpaceX says that this government-focused service will require “Starshield user equipment,” which likely resembles Starlink’s in operation but meets certain special standards of ruggedness, access, documentation and compatibility with existing networks and assets. As Starlink itself has multiple tiers of ground station, from ordinary consumer rooftop type to paired extra-tough nautical type, the Starshield version is probably going to be one of the high-end ones, souped up (for instance with “additional high-assurance cryptographic capability”) and with a price tag to match.

More importantly this move helps separate government work from consumer work. The company has lamented that its deployment of thousands of terminals in Ukraine has resulted in a quagmire of legal and financial finger-pointing: Ukraine can’t pay, its allies didn’t agree to pay and SpaceX can’t provide the expensive service for free indefinitely. This is partly because the whole network was really never meant to be used in this fashion, and grafting a military/aid operation onto a consumer product has led to unforeseen consequences.

By being more intentional about what services it provides government entities and under what terms, SpaceX probably hopes to avoid the blurred lines between being a global broadband provider and being a supplier of military intelligence. Both may be very profitable in their way, but rarely does one product serve both purposes adequately.

Image Credits: SpaceX

The company also claims to produce a modular satellite bus for diverse mission types, though again this is unsubstantiated — not to say it isn’t so, but the capability is simply stated, not shown with anything beyond a wireframe image.

Exactly how far SpaceX has gotten in achieving the capabilities it describes here is anyone’s guess — they may well have prototyped some of these things out already with some prospective customers, or this may just be a statement of intent with those customers in mind. Regardless of which is the case just now, it seems clear we will be hearing more about this service as its roles escape attempts at secrecy — for example, it is difficult to launch a major Earth observation satellite without anyone knowing.

I have asked SpaceX for more information on its customers and capabilities and will update this post if the company responds.

SpaceX goes full defense contractor with national security-focused Starshield by Devin Coldewey originally published on TechCrunch

Daily Crunch: Another Salesforce C-suite exit — Slack CEO Stewart Butterfield will step down in January

To get a roundup of TechCrunch’s biggest and most important stories delivered to your inbox every day at 3 p.m. PDT, subscribe here.

Obviously you love the Daily Crunch, but did you know that we’ve got a whole lineup of truly amazing newsletters from across the site? Sarah’s This Week in Apps is consistently insightful and interesting about what happens on our handheld supercomputers. The Interchange is Mary Ann’s deep dive into the world of fintech startups, and The Station is the summary of everything transportation, lovingly assembled by the incomparable Kirsten. Finally, Greg’s Week in Review is the “oh crap, I didn’t have time to read TechCrunch this week” summary so you don’t make a complete fool of yourself at the watercooler when Monday rolls around.

We also have space newsletters, robotics newsletters, and more. Check out the full list, and go on a subscribing spree. Go on, what else is there to do in December? — Christine and Haje

The TechCrunch Top 3

Ch-ch-ch-changes: It’s another C-suite shake-up over at Salesforce. Ron reports that Slack CEO Stewart Butterfield, who joined Salesforce when the company acquired Slack in 2020, says he is stepping down in January. This news follows some other recent out-of-the-blue departures in the past week, including that of Salesforce co-CEO Bret Taylor, who announced his resignation.
A different take on social commerce: Much of the livestream commerce we see today has to do with buying clothes or some fancy vintage playing card. However, Kapu is putting an interesting spin on it, coming out of stealth with $8 million in new funding to create a social commerce site that helps Kenyans find groceries at lower prices. Annie has more.
AI is our life now: We are now using artificial intelligence to create our own avatars and write letters about potential layoffs. Twelve Labs’ founder Jae Lee believes this should be extended to searchable video, so that’s what the company set out to build. Today, it announced $12 million in fresh capital to continue developing AI that understands the context of videos, Kyle writes.

Startups and VC

Loft Dynamics, a Swiss startup creating virtual reality simulation technology for helicopter pilots, has raised $20 million in funding from U.S. venture capital firms, including Craft Ventures, Sky Dayton and Up Ventures, Paul reports. Alongside the funding, the company is also formally announcing its name change from VRM Switzerland to Loft Dynamics.

Apropos VR…demand for real estate VR booms, reports Mike, writing that Founders Fund leads a $16 million round into the Giraffe360 platform.

Here’s five more news story and an invite to the consumer electronics startups out there:

From nonduality to nonexistence: Romain writes that Zenly was the best social app and that he’s sad it’ll shutter on February 3.
Plus d’Euros: French VC firm Partech has closed its fourth fund, a €120 million ($124 million) seed fund, Romain reports.
Row, row, row your code: Haje writes that if Rowy has its way, if you can use Excel, you can build software.
Premature expiration: Devin reports that it looks like sex tech startup Lora DiCarlo is done for.
Today’s special is…: Aigens scales digital menus in Asia, reports Rita, with a $14 million round led by Ant.
Coming to CES? Haje, Brian, and a few of the other TC crew are going, and we want to meet your startup in Vegas in January, writes Brian.

Which way is up? The end of free money and the importance of keeping cash on hand

Image Credits: PM Images (opens in a new window) / Getty Images

In simpler times, founders could often satisfy investors just by showing how quickly their company was meeting growth expectations.

“Well, investors today care about the less-distant future,” said Max Schireson, an operating partner at Battery Ventures.

“They care about how much money they need to put into your company to get to that future and when it will arrive.” In a guest post for TC+, he shares frank advice and multiple scenarios that can help founders meet investor expectations during tough times.

“They say time is the one thing you can’t buy, but in fact, time is the easiest thing to buy at a startup.”

Three more from the TC+ team:

For goodness SaaS: Kyle writes how companies can slash ballooning SaaS costs.
A boom in doom and gloom: Tim argues that climate tech is not doomed, despite climate doom.
Slowing the Amazon River: Ron wonders if the era of constant innovation at Amazon is over.

TechCrunch+ is our membership program that helps founders and startup teams get ahead of the pack. You can sign up here. Use code “DC” for a 15% discount on an annual subscription!

Big Tech Inc.

Facebook wants to know your age, but before you clutch your pearls, it’s actually for a good reason. Meta is bringing its age verification technology into its Facebook Dating product, Sarah reports. We already don’t know, or want to know, how 50-year-olds meet 16-year-olds, but Meta definitely doesn’t want to gain a reputation for being THAT place.

And now, we have five more for you:

“No” to jail, for now: Kate reports that a Korean court rejected warrants to detain eight former Terraform Labs employees and investors related to the Luna collapse. The court said the people had a right to fight their cases.
Some people think this guy belongs in jail, too: It’s unlikely that Sam Bankman-Fried will testify in a House Committee hearing on the FTX collapse, taking place on December 13, Manish writes.
Feature drop: Google Pixel users, like Christine, rejoice! An update has new features like clear calling and Google One VPN, Ivan writes.
Going international: General Motors’ BrightDrop e-delivery van business found a new home in Canada with DHL Express, Rebecca reports.
There goes the neighborhood: Elon Musk has been busy bringing back an infamous neo-Nazi on Twitter and vicariously publishing internal emails related to the Hunter Biden laptop drama. Taylor has more on both of these stories.

Daily Crunch: Another Salesforce C-suite exit — Slack CEO Stewart Butterfield will step down in January by Christine Hall originally published on TechCrunch

Investors sound the alarm about possible private equity tech deals

Enterprise spend management software company Coupa’s investors are ringing the alarm about a possible sale to private equity, concerned that any such transaction in the current investment environment may unreasonably limit its value.

This is something you don’t see every day. Last week, rumors surfaced that Vista Equity Partners was interested in buying Coupa. Today, Coupa’s largest shareholder, HMI Capital, with 4.8% of the stock, made a letter to the Coupa board public, stating that it would oppose any deal that it believed undervalued the company.

It pegged its value at a minimum of $95 per share. Coupa’s share price hovered around $63 this afternoon, down nearly 3%. Like many SaaS stocks, Coupa’s value has dropped precipitously in 2022, down 60% for the year.

HMI is worried that Coupa’s current value doesn’t reflect what it believes will be a lucrative future once it gets beyond today’s troubled valuation market for software companies.

Unlike the poison-pen letters we are used to seeing from activist investors, this one was full of nothing but praise for the company (probably because its goals were very different):

As we have conveyed to the Board and management team, we invested in Coupa based on our belief that it is an excellent business. Its management team — most notably Chief Executive Officer Rob Bernshteyn — has done an exceptional job building the Company from a startup to a clear market leader, while establishing business spend management as its own critical category. Perhaps most importantly, Coupa’s outlook for future growth and long-term value creation is bright. As Mr. Bernshteyn himself stated only a few months ago:

“Now near-term scenario aside, we are proudly the clear leader in business spend management. Our total addressable market is massive and under-penetrated and we are excited as ever in our pursuit to revolutionize this market and deliver customer success like never seen before.”

It’s our view that we speak for many other shareholders when we say that we would be pleased to own Coupa for the foreseeable future and to bet on the team to continue to build momentum and execute its proven strategy.

It’s almost odd to see an external investor praising a company’s leadership, as most public investor comment is negative. Here we do not see an investor begging management to cut costs or change direction. Instead, it’s praise and a belief in greater value yet to come.

The sunny view of Coupa’s future conflicts with present public market sentiment about the future value of tech companies. That prevailing viewpoint, crossed with huge amounts of private equity dry powder, may have put PE investors in a deal-making mood.

HMI is begging the Coupa crew to stick it out, or at least demand more than they otherwise might settle for. Of course, this is HMI talking its own book, but there may be some substance to its argument. Let’s check the math — and a comp.

Investors sound the alarm about possible private equity tech deals by Ron Miller originally published on TechCrunch

Gift Guide: More than 20 STEM gift ideas to inspire kids to code

All aboard the learn-to-code train! As another holiday season looms and kids everywhere start clamouring for shiny stuff, we’ve got you covered with a bumper edition of our annual STEM gift guide. This year’s guide is packed with more than 20 ideas to engage toddlers, excite tweens and inspire teens to get under the hood of the tech world through play.

Products in the STEM learning category typically promise to give kids a head-start on learning core coding and/or electronics concepts while they play. How much learning actually gets done is still a bit of an open questio but ideas in the category have generally converged around a few approaches — so there’s a stronger sense of ‘tried & tested’ out there now. We’re also encouraged to see increasing attention to diversity in STEM.

A few other trends are evident this year — most notably around supply issues linked to the pandemic-triggered (and still lingering) global supply chain crunch. Said crunch appears to be having some impact on product choice and availability.

One category-adjacent example: Sphero’s RVR+, a (mostly) school and makerspace focused programmable robot, was listed as unavailable for supply reasons when we started compiling the guide earlier this fall — but at the time of writing the website reports it’s back in stock “just in time for the holidays.”

‘Out of stock’ labels also remain a regular sight on STEM-focused marketplaces that range maker hardware, such as AdaFruit and Pimoroni. The latter told us recently it’s been managing ongoing shortages of Raspberry Pi hardware by rejigging featured products to mesh with available inventory. So creative stock hacks abound.

Supply issues may also be further decelerating innovation in what had already become a less experimental market compared to earlier, investor-fuelled boom years. The vertical had relatively few wholly new product launches for 2022, as players focused on shifting existing inventory instead.

The upshot is a feeling there’s a bit less fun and excitement to go around than in years past, furthered as more STEM startups pivoted to selling to the formal educational market (not so much to home buyers).

That crossover, from clever toys nudging informal learning to curriculum-focused products catering to formal education, is reshaping the kind of STEM kit on offer and making lots of stuff in the category feel, well, more serious. But — don’t despair — there’s still plenty of fun choices for gift-buying parents and relatives wanting to give budding builders and tiny techies a leg up to grasp the power and possibility of programming.

Suggested gifts in our guide run the gamut from a doll that sings the praises of coding to line-following programmable robots and hardware hacking electronics maker kits. There’s also a DIY computer kit and a sensor-packed robot-on-wheels that can dance and talk back.

As ever, we’ve made sure to include gift options that hit a wide range of price-points, starting at just $6 for a bare microcontroller that throws your kid in at the STEM learning deep end (they’ll need to source other electronics components, have access to a computer and the Internet, and display no shortage of determination to unlock the learning potential with that one), scaling all the way up to $250 for a talking robot on wheels that’s packed with (perhaps too many) distracting bells & whistles, with plenty of choice and price-tags in between. Enjoy!

Gift suggestions are grouped by age for easy reference: 2+, 5+, 8+ and 12+ years.

This article contains links to affiliate partners where available. When you buy through these links, TechCrunch may earn an affiliate commission.

2+ years

Lego Coding Express

Image Credits: Lego

Lego’s Coding Express is a train set for playful early years coding exploration — featuring 200+ Duplo bricks, including a push-and-go motorized train with lights, sounds and color sensors. The sensors interact with colored action coding blocks which kids place on the track. The set includes activity and inspiration cards to get your toddler started on grasping linear logic — with a free app to expand the parentally guided learning potential.

Age: 2-5
Price: $230from Amazon
Made by: Lego Education

Codie the Coder

Image Credits: Surprise Powerz

Surprise Powerz, the U.S.-based startup behind this cute learning doll, Codie the Coder, was set up back in August 2019 with a STEM-centric mission to inspire girls, especially black and Latina girls, to “break barriers, solve problems and have fun doing it”. “We aim to get more STEM role models in the hands of more girls all across the nation to build their confidence in these critical subjects early,” it explains.

Founder and CEO, Kristel Bell, tells TechCrunch that Codie the learning doll speaks over 75 “coding-related phrases” — just squeeze her hand to hear the doll speak short phrases and questions. She says the goal is to inspire little girls to learn “they also have a place within the tech space” by seeing and hearing a coder role model which looks and talks like them, and engages them in an open ended conversation as they play. “Representation matters,” she adds.

Surprise Powerz’ STEM learning dolls are designed for pre-schoolers and up. Absolutely no screens required.

Age: 2-5
Price: $65 from Surprise Powerz
Made by: Surprise Powerz

Botzees Mini

Image Credits: Pai Technology/Botzees Toys

Botzees’ expressive, line-sensing bot for budding programmers offers a variety of ways for kids to learn core coding concepts as they play. As well as programming challenges, the Botzees Mini pack features math and music puzzles too, throwing a little melody into the tech mix.

Kids control the bot by placing it on the pre-drawn path on maps provided — or by line drawing themselves. The coding element involves kids placing command cards on the robot’s path to control its motion, sound and appearance, with feedback coming at them via the bot’s changing facial expressions (which are displayed on its built-in screen). Supported functions cover programming basics like cause and effect, if/then logic, sequencing and debugging. No additional screens or companion apps are required.

Age: 3+
Price: $70from Amazon
Made by: Pai Technology/Botzees Toys

Switcheroo Coding Crew

Image Credits: Learning Resources

Turn learning STEM basics into an “interactive coding adventure” with Switcheroo Coding Crew. The kit aims to teach kids basic coding concepts, early counting and math skills, critical thinking, spatial concepts, sequential logic and more with the help of a battery-powered programmable truck that sports changeable colored shells.

The play-set includes a town-themed puzzle board and a set of mission challenge cards to get things rolling. Screen-free play.

Age: 4-7
Price: $30 from Amazon
Made by: Learning Resources

5+ years

codeSpark Academy – gift subscription

Image Credits: codeSpark Academy

For kids with access to an iPad (or Android-powered tablet), codeSpark Academy sells a range of subscription plans for ‘learn to code’ content in a gaming wrapper which aims to foster sequencing skills and help kids get a head start in STEM as they design and play. The software is centered on a proprietary game-making platform that features a set of cartoon characters called “The Foos.” Kids learn problem solving and coding by engaging with challenges and building their own games.

CodeSpark Academy says its approach is based on research-backed curriculum from MIT and Princeton. Its system includes a word-free interface to maximize accessibility and self-directed learning, with a claim that no prior experience is needed for kids to get coding.

Gift subscriptions start at around $60 per 6 months access — though you’ll have to stump up more to also get the ‘free’ plush toy pictured above…

Ages: 5-9
Price: Gift plans from $60 for 6 months (up to $225 for a lifetime sub, ‘Glitch’ plushy included)
Made by: codeSpark

Miko 3

Image Credits: Miko

Meet Miko 3: A sensor-laden, emoting tablet-on-wheels that’s been programmed to entertain and inform curious young minds. Encased in a brightly colored, robot-shaped shell is a voice- and face-recognizing AI (plus built-in speaker) which enables it to respond to questions and commands, piping up with interesting facts or playing tunes for a child’s dance party, and so on. A lot of what Miko offers is pure entertainment so there’s a risk STEM learning takes a back seat to other more fun interactions. This is also definitely not a tech lite (nor screen-free) learning option.

Still, parents hoping to accelerate the development of budding young coders might be won over by educational content the bot can put on tap for kids — via its curated ‘Talents’ app store. Available content includes Kidloland’s Coding School app. Again, though, good luck getting children to focus on something that techie when they could just be asking the bot to spin circles or play hide and seek…

Some of Mike 3’s digital content is bundled with the base price of the bot; unlimited premium kids content requires a subscription.

Ages: 5-10
Price: $249 (plus optional subscription)
Made by: Miko

Osmo Coding Starter Kit

Image Credits: Osmo

If your wannabe coder already has access to a tablet like an iPad, Osmo’s clever add-on system turns screen time into STEM learning time via hybrid digital/physical play. The product is comprised of a base stand to lift the screen off the table, making room for little hands to physically build lines of code, and a reflector that enables the software to keep track of their table-top programming.

As the name suggests, the Osmo Coding Starter Kit is the place to start. The kit is designed to help kids build coding skills in progression — walking them through three hands-on learning games. Kids learn coding basics by putting together the colorful physical programming blocks, and get real-time feedback by watching their coding adventure play out on the screen.

Ages: 5-10
Price: $99
Made by: Osmo

Tacto Coding

Image Credits: PlayShifu

For another spin on physical/virtual play, PlayShifu’s Tacto Coding system turns an existing (compatible) tablet into a dedicated kids coding station. Children are encouraged to get to grips with programming basics by playing through a variety of mini challenges with the look and feel of puzzle games.

The attention-grabbing twist is they interact with digital content on the touchscreen using a range of (physical) figurines arrayed around the tablet base-station. This (er, metaverse ready?) cross-over device is aimed at turning learning coding into child’s play — and a physical-virtual adventure.

PlayShifu’s child-friendly learn-to-code approach also has advanced modes to support kids as they progress, including offering Scratch programming language support so they can try their hand at block-based coding.

Age: 6-12
Price: $40 from Amazon
Made by: PlayShifu

8+ years

Ozobot STEAM Kits

Image Credits: Ozobot

Ozobot has been selling mini programmable STEM learning robots for several years, but these days it’s a lot more focused on the K-12 educational market. New for 2022 it’s added a bunch of STEAM kits into its product mix. You’ll need to either already own or buy one (or more) Ozobots to get the most out of these — but if you already have Ozobot kit at home they could be worth a look.

The OzoGoes to the Solar System kit combines the pull of its rolling robots with a paper-based educational kit for learning about the planets in our solar system, with the bots helping to illustrate their movements. Another kit — OzoGoes to the Sun, Earth & Moon — is similar but doesn’t require as many Ozobots to get the kinetic learning rolling.OzoGoes Around A Sundial (pictured above) has kids deploying their bot as a tool to help learn how to use the sun to tell the time.

Ozobot says the STEAM learning kits were designed by “experienced educators and curriculum developers” so homeschooling parents especially may want to check them out. The kits feature both screen-free (paper-based) and on-screen project-based activities.

Age: 8+
Price: From $30
Made by: Ozobot

Sphero Mini Activity Kit – Kids Coding Activity Kit

Image Credits: Sphero

Another programmable robotics maker that’s been rolling around the STEM learning space for years is Sphero. It’s also become increasingly geared towards the school markets. But a solid entry point for parents looking to give their kids a very-high energy taste of making code approachable (and fun) is the Sphero Mini Activity Kit – Kids Coding Activity Kit.

The kit comes with a Sphero Mini programmable bot plus a construction set containing STEM-inspired activity cards and bits and bobs so kids can build mazes, design obstacle courses, build towers, play croquet and more. The Sphero bot can be code-controlled using Blocks drive mode — which relies on basic drag-and-drop coding commands — or kids can tap into the Sphero Edu app to expand their programming knowledge. There’s also a Sphero Play app that turns the spherical gizmo into a game controller for arcade-style games if/when they get bored of all the STEM activities.

Age: 8+
Price: $89
Made by: Sphero

littleBits Electronic Music Inventor Kit

Image Credits: Sphero

The littleBits STEM learning brand — which, since 2019, has been owned by the aforementioned Sphero — is fixed on helping kids learn how electronics hardware works, one connectable component at a time. Perfect if you’ve got a budding hardware hacker on your hands.

Musically minded kids should appreciate putting together (and rocking out) with this littleBits Electronic Music Inventor Kit, which guides them to build a “rockin’ synth guitar,” air drums or another unique creation of their own using the 8 bits and 18 accessories provided in the kit.

Age: 8+
Price: $100
Made by: Sphero

Snap Circuits Code Journey

Image Credits: Elenco

Veteran electrical engineering STEM brand Elenco has devised a kid-friendly system of encased, snap together electronics components to encourage children to learn about circuits by building them. The Code Journey rover kit, pictured above, adds programming into this mix.

Children are encouraged to get coding by assembling circuits that they mount on the Snap Rover — using components like LEDs, a speaker, jump wires and more — and then by coding commands to send to the vehicle so it can take instructions and perform missions. Programming the rover is done using a simple graphical interface in the Snap Circuits App. There’s also support for Blockly Coding’s drag-and-drop interface as kids’ skills advance.

Age: 8+
Price: $103.74
Made by: Elenco

Coding Lab GinoBot

Image Credits: Engino

Cyprus-based Engino has spent years developing a programmable robotics platform for teaching STEM and physical robotics. Newly introduced into the US and UK markets is its Coding Lab: GinoBot. The programmable 4-wheel drive robotics vehicle (which can be assembled into one of two models) packs Wi-Fi and Bluetooth connectivity, color sensors, programmable RGB lights, an ultrasound sensor and infrared distance sensors — for powerful programmable play.

Also on board: Built-in connectors to enable structural and mechanical expansion, and the ability to plug in open electronics (such as Raspberry Pi, Microbit and Arduino, or Engino’s own ERP sensor modules) to further expand what’s on board, paving the way for a budding hardware hacker to grow their engineering skills.

Age: 9+
Price: $137.23
Made by: Engino

Code Rocket

Image Credits: Let’s Start Coding

Stuff a stocking with Code Rocket and spread a little STEM smarts! This dinky, spaceship-shaped circuit board is armed with built-in LED lights, speaker, and buttons, and designed to teach C++ coding through hands-on electronics. As well as space-themed electronics, the pack provides access to over 20 online lessons. Said lessons introduce kids to concepts like loops, variables and functions as they program the rocket booster LEDs, code a countdown to liftoff, send morse code sound messages, and more. Both block-based and typed C++ coding is supported. Access to a desktop computer to program the electronics is required.

Age: 8-12
Price: $54.99
Made by:Let’s Start Coding

Piper Make Light Show

Image Credits: Piper

San Francisco-based Piper Make has clocked up almost a decade in the STEM learning gizmo business — expanding to offer a range of products that encourage hands-on learning and typically blend electronics hardware with distinctive laser-etched wood base structures. The Piper Make Light Show, pictured above, offers a festive intro to programmable electronics. Curious youngsters will be encouraged to learn how to program the light panel to do stuff like display scrolling messages, display light art or host a Tetris-style game.

Programming the kit requires access to a computer (using the bundled micro-USB to USB-A cable) to access Piper’s drag-and-drop coding platform: make.playpiper.com. For the Light Show to work you’ll also need to have Piper’s base kit (aka, the Piper Make Starter Kit) — which comes with a Raspberry Pi microcontroller that’s used to power all its products.

Age: 8+
Price: $59 for the Make Light Show (plus $49 for a Starter Kit if required)
Made by: Piper

imagiCharm besties starter kit

Image Credits: imagiLabs

Got twin girls with an eye for tech? Get them into coding with imagiCharm’s besties starter kit — which ships a pair of its customizable tween-friendly, Internet connected “charms”. The Swedish startup’s Tamagotchi-esque connected gizmos are designed to inspire by teaching how to use code to change the look and function of the 8×8 matrix of colored LED lights.

The battery-powered charms can be attached to a school backpack or worn as digital jewellery to show off its coding owner’s latest look/vibe/mood/feels.

Age:8-13+
Price:$136
Made by:imagi Labs

Smart Coding Watch Kit

Image Credits: Elecfreaks

Has your little engineer been eyeing up the Apple Watch? Why not get them to think outside the hermetically sealed, high shine, proprietary tech box by putting together their own smart watch? And learning to program it on open source hardware with this micro:bit add-on Smart Coding Watch Kit. The STEM learning kit guides them through assembling the wearable and learning how to program functions for it by connecting it to a computer and using free programming software. Options include graphical programming languages like Makecode and Scratch, or microPython or JavaScript.

(As well as access to a desktop computer for programming the hardware, you’ll need a micro:bit board — not included in the watch kit — to power the wearable. If you don’t already own that STEM-focused microprocessor remember to pick one of those up too.)

Age: 8+
Price: $9.90 (+ $29.99 for a micro:bit)
Made by: Elecfreaks

Boolean Box

Image Credits: Boolean Girl Tech

How about getting your kid learning programming by building a computer of their own to code on? That’s the goal of build-it-yourself Boolean Box.

The DIY computer kit (powered by a Raspberry Pi 4) has been around for a few years — and was conceived with the goal of encouraging girls into coding. But it’s also designed to be an approachable option for children of any gender to up their STEM smarts.

Once kids have built and booted up the computer they’ll be nudged to start coding. Whether by learning to program mini games with Scratch, code Minecraft hacks with Python, or play at being a budding inventor using the bundled breadboard and electronics components to do circuit projecting… just add a TV or monitor and they’re good to go.

Age: 8+
Price:
$199.99
Made by: Boolean Girl Tech

12+ years

CircuitPython Starter Kit with AdaFruit Itsy Bitsy M4

Image Credits: AdaFruit

Get your little maker started as a hardware hobbyist with CircuitPython using this cute electronics Starter Kit from AdaFruit — powered by an ItsyBitsy M4 microprocessor.

AdaFruit’s website offers tons of beginner guides for help learning all the ins-and-outs of electronics and hardware making. (Note: As well as containing small parts, some light soldering is required to attach headers to the ItsyBitsy, so the build process may require parental supervision.)

Age: Young engineer
Price: $24.95
Made by: AdaFruit

Legendary

Image Credits: Represented Collective

Give the gift of inspiration and a sense of belonging to the next generation of expectation-smashing women in STEM. Legendary, by Represented Collective, is a 56-card collection — featuring 41 profiles of professional women in science, technology, engineering and math — curated to showcase the often underrepresented contributions of primarily Black, Indigenous, Women of Color (BIWoC), and to celebrate what it describes as “a diverse set of women and their accomplished, yet nuanced stories”.

Age: Precocious tween+
Price:
$40
Made by: Represented Collective

Trilobot Complete Kit

Image Credits: Pimoroni

Pimoroni is ranging this handsomely rugged-looking, Raspberry Pi-based Trilobot Complete Kit, which it bills as a “mid-level” robot learning platform.

Designed with educational use in mind, kids first assemble the bot then boot up the on-board Pi and install the Trilobot Python library in order to get tinkering with programming by running examples that let them play around with the bot’s features and functions. As their mastery grows, there’s a possibility of connecting a controller to remote control the rover — if they can figure out the code calls.

The three-wheeled bot sports a pair of grippy, “moon-buggy” wheels up front and a ball caster round the back for 2x front wheel drive. As well as its Pi microprocessor engine, the bot is armed with a Pi v2 camera and an ultrasonic distance sensor. It also sports RGB LED underlighting to dial up the visual fun. Plus you get a 32GB microSD card and 5000mAh USB-C battery pack in this version of the kit (so check local shipping rules to make sure you’re able to receive the battery pack).

Pimoroni has created a detailed, step-by-step beginner tutorial for assembling the Trilobot. After that it’s Python code tinkering time…

Age: Roboticist in training
Price: $120.16
Made by: Pimoroni

Raspberry Pi Pico w

Image Credits: Raspberry Pi

Don’t overlook the gift of pure, unadulterated Raspberry Pi. It’s an incredibly budget-friendly option for STEM learning — such as the $6 Pi Pico microcontroller with wi-fi, pictured above, which is great for powering DIY electro-mechanical projects. Though you will need a bunch of other components and access to a computer to program the hardware.

As we reported earlier this fall, Pi production has been hit by ongoing supply chain challenges since the pandemic but the Pico line is less affected than some of its other (bigger) boards so should be easier to track down.

Pure Pi comes with the baked in advantage of encouraging kids to seek out resources to help themselves learn which was of course the original philosophy behind the wildly successful maker hardware. The Raspberry Pi Foundation offers a solid starting point for setting up newly acquired Pi — and taking the first steps in a learn-to-code journey.

They’ve also published an official guide for MicroPython on the Pico in handy paper-based book form which could make a nice additional gift.

Age: It depends
Price: $6
Made by: Raspberry Pi

Gift Guide: More than 20 STEM gift ideas to inspire kids to code by Natasha Lomas originally published on TechCrunch

Pin It on Pinterest