A large majority of Americans feel that religion is losing influence in public life, according to a 2019 Pew Research Center survey. While some say this is a good thing, many more view it as a negative development, reflecting the broad tendency of Americans to see religion as a positive force in society.