It’s always better

to tell the truth.

I’ve heard this

over & over

from every adult

ever.


But are they right?


What about when

the truth is hard to hear?

When it’s ugly & dark?

When it hurts?

When it makes everyone sad?


Is it ever better

to just

stay silent?


Or will the truth

actually always

set you free?