Page 1 of 1

"Can Books Really Change Our Perspective on Humanity?"

Posted: 12 Jun 2025, 11:26
by Kunal Padwal
Absolutely. I recently read a book called *Kenny* by Mark Carew, and it genuinely made me question how we treat our planet and each other. The story is told through a father explaining to his son why aliens might not want to contact humans — and the reasons he gives are painfully believable. It really shifted my thinking.

It’s interesting how fiction can hold up a mirror to reality. Books like this don't just entertain — they force us to pause, reflect, and sometimes even change a bit from the inside.

Has anyone else read a book that made them rethink something about humanity or themselves?

Re: "Can Books Really Change Our Perspective on Humanity?"

Posted: 12 Jun 2025, 11:36
by Kunal Padwal
Absolutely yes. Books have the power to completely change how we see the world and ourselves. I recently read *Kenny* by Mark Carew, and it left a surprisingly deep impact on me. The story is about a father explaining to his son why aliens may choose not to contact humans — and the reasoning feels emotional and disturbingly accurate.

It made me reflect on how we treat our planet, how we treat one another, and whether humanity is really worthy of being part of something bigger. Fiction like this doesn’t just entertain — it opens your mind to uncomfortable truths.

Have you ever read a book that made you stop and seriously reflect on human nature? I’d love to hear your thoughts or any recommendations that challenged your worldview.

Re: "Can Books Really Change Our Perspective on Humanity?"

Posted: 12 Jun 2025, 11:40
by Kunal Padwal
Absolutely yes. Books have the power to completely change how we see the world and ourselves. I recently read *Kenny* by Mark Carew, and it left a surprisingly deep impact on me. The story is about a father explaining to his son why aliens may choose not to contact humans — and the reasoning feels emotional and disturbingly accurate.

It made me reflect on how we treat our planet, how we treat one another, and whether humanity is really worthy of being part of something bigger. Fiction like this doesn’t just entertain — it opens your mind to uncomfortable truths.

Have you ever read a book that made you stop and seriously reflect on human nature? I’d love to hear your thoughts or any recommendations that challenged your worldview.