I like it, but I don't like seeing gay men having sex or the showing of Muslim people when so many are causing so many problems in our country. Why does Hollywood give in to these two groups and show this in every show? I know it exists, but am tired of seeing it in every show! I'm not the only one who feels this way. Probably won't watch anymore. It's a shame, it could have been a good show.