The NFL's Philadelphia Eagles are entering a multi-year partnership with LifeBrand, a startup that specializes in social media background checks for personal and corporate brands. 

LifeBrand uses AI-powered technology to perform "social media health checks" aimed at detecting and deleting posts that are potentially harmful to brands. The startup is headquartered not far from the City of Brotherly Love in West Chester, Pennsylvania. 

"As social media becomes a big part of our lives and influences our day-to-day lives, I think the ability to do a little bit of a social media health check on your own social accounts can really help and quickly detect anything that might be potentially offensive or harmful," Catherine Carlson, senior vice president of revenue and strategy for the team, told Cheddar.

The agreement will begin with the NFL Draft. LifeBrand will receive naming rights to the Eagle Draft Room at the NovaCare Complex and the Journey to the Draft podcast. Lincoln Financial Field's Northeast Gate will also be renamed to LifeBrand gate.  

Will the team use the technology on its own players? Carlson said that's still being worked out, but that employees and players in the franchise could probably benefit from a social media checkup. 

"If you think about it, our players have a sizable social media footprint, as does the Eagles, and with that comes a really huge opportunity and responsibility," she said. 

Fans will have the opportunity to use the technology as well. 

"We'll also give fans the opportunity to do on-site, quick health scans of their social media accounts as they're coming to the stadium," Carlson said. 

The process is voluntary and quick, but the city's famously prickly sports fans might not willingly submit their social media feeds to scrutiny.  

Share:
More In Business
Sex is a big market for the AI industry. ChatGPT won’t be the first to try to profit from it
OpenAI has announced that ChatGPT will soon engage in "erotica for verified adults." CEO Sam Altman says the company aims to allow more user freedom for adults while setting limits for teens. OpenAI isn't the first to explore sexualized AI, but previous attempts have faced legal and societal challenges. Altman believes OpenAI isn't the "moral police" and wants to differentiate content similar to how Hollywood differentiates R-rated movies. This move could help OpenAI, which is losing money, turn a profit. However, experts express concerns about the impact on real-world relationships and the potential for misuse.
Load More