It’s always better
to tell the truth.
I’ve heard this
over & over
from every adult
ever.
But are they right?
What about when
the truth is hard to hear?
When it’s ugly & dark?
When it hurts?
When it makes everyone sad?
Is it ever better
to just
stay silent?
Or will the truth
actually always
set you free?