There are a lot of movies out there that I would hate to be paid to do, some real demeaning, real woman-denigrating stuff. It is up to women to change their roles. They are going to have to write the stuff and do it. And they will.
Both class and race survive education, and neither should. What is education then? If it doesn’t help a human being to recognize that humanity is humanity, what is it for? So you can make a bigger salary than other people?