We are independent & ad-supported. We may earn a commission for purchases made through our links.
Advertiser Disclosure
Our website is an independent, advertising-supported platform. We provide our content free of charge to our readers, and to keep it that way, we rely on revenue generated through advertisements and affiliate partnerships. This means that when you click on certain links on our site and make a purchase, we may earn a commission. Learn more.
How We Make Money
We sustain our operations through affiliate commissions and advertising. If you click on an affiliate link and make a purchase, we may receive a commission from the merchant at no additional cost to you. We also display advertisements on our website, which help generate revenue to support our work and keep our content free for readers. Our editorial team operates independently of our advertising and affiliate partnerships to ensure that our content remains unbiased and focused on providing you with the best information and recommendations based on thorough research and honest evaluations. To remain transparent, we’ve provided a list of our current affiliate partners here.

Our Promise to you

Founded in 2002, our company has been a trusted resource for readers seeking informative and engaging content. Our dedication to quality remains unwavering—and will never change. We follow a strict editorial policy, ensuring that our content is authored by highly qualified professionals and edited by subject matter experts. This guarantees that everything we publish is objective, accurate, and trustworthy.

Over the years, we've refined our approach to cover a wide range of topics, providing readers with reliable and practical advice to enhance their knowledge and skills. That's why millions of readers turn to us each year. Join us in celebrating the joy of learning, guided by standards you can trust.

What is Entropy?

By S. Mithra
Updated: May 21, 2024

Entropy describes the tendency for systems to go from a state of higher organization to a state of lowest organization on a molecular level. In your day-to-day life, you intuitively understand how entropy works whenever you pour sugar in your coffee or melt an ice cube in a glass. Entropy can affect the space into which a substance spreads, its phase change from solid to liquid to gas, or its position. In physics, entropy is a mathematical measurement of a change from greater to lesser potential energy, related to the second law of thermodynamics.

Entropy comes from a Greek word meaning, "transformation." This definition gives us insight into why things seemingly transform for no reason. Systems can only maintain organization on a molecular level as long as energy is added. For example, water will boil only as long as you hold a pan over flames. You're adding heat, a form of kinetic energy, to speed up the molecules in the water. If the heat source is removed, we all can guess that the water will gradually cool to about room temperature. This is due to entropy, because the water molecules tend to use up their accumulated potential energy, release heat, and end up with a lower potential energy.

Temperature isn't the only transformation involved in entropy. The changes always involve moving from disequilibrium to equilibrium, consistent with moving to decreasing order. For instance, molecules always spread out to uniformly fill a container. When we drip food coloring in a clear glass of water, even if we don't stir it, that united concentration of a drop will gradually spread out until every part of the water has the same density of color.

Another type of entropy that has to do with visible motion (as opposed to the invisible motion of heat) involves gravity. Unless we put energy into a system, like an arm and a ball, by holding up an object, it falls toward the ground. An elevated position has higher potential energy. It gets converted into kinetic energy of motion as the object falls. The object always ends up with the position of lowest possible potential energy, such as resting against the floor.

In more technical terms, entropy is a specific value that measures how much energy is released in a system when it settles into the lowest potential energy. Entropy assesses the amount of disorder, understood as a change in heat, from an earlier point to a later point in time. This must happen in a "closed" system, where no energy leaks in or out. Theoretically, that can be measured, but practically it is very difficult to create an absolutely closed scenario. In the food coloring example given above, some of the food coloring solution might be evaporating, a separate process from the uniform distribution of a solute.

All The Science is dedicated to providing accurate and trustworthy information. We carefully select reputable sources and employ a rigorous fact-checking process to maintain the highest standards. To learn more about our commitment to accuracy, read our editorial process.
Discussion Comments
By anon114831 — On Sep 30, 2010

No the opposite. Higher entropy means less chaos (energy in the system has dissipated, we are close to the death of the universe, close to equilibrium). Low entropy i.e. when we add more energy to the system it gets more chaotic.

From an information theory point of view, "entropy" is how much information the message contains. Higher entropy means the message contain more information and vice versa.

When thinking about this you have to take into account whether you are talking about the sender or the receiver of the message.

For example, let's say the sender sends the message with lots of information (high entropy from sender's point of view). Now the receiver gets the message. Does this made the things clearer (the message had high entropy to receive.) or more ambiguous (the message had low entropy to receive).

By anon106838 — On Aug 27, 2010

could you do a definition from the information theory point of view? that would be cool.

By anon90378 — On Jun 15, 2010

If you think of a closed system, as points with different pressures, they will eventually diffuse into a equilibrium with a single pressure over all points. This is the same as going from having potential energy to having none! Thus the system will see no motion, and you won't be able to provoke it unless you added more energy.

In the context of the enormous universe, the same force are acting upon it will eventual cancel out any differences and densities between the various forms of energy. The end state will be the hypothetical heat death of the universe.

"Entropy is a mathematical measurement of a change from greater to lesser potential energy"

is a very good definition. Check the potential energy link if you still don't get entropy.

By anon83176 — On May 09, 2010

is this really that hard to understand? I'm in seventh grade doing a project on thermodynamics.

I have a good explanation. when japan put a crapload of energy (bombs) into pearl harbor, how come the energy didn't clean up the site instead of blowing it to bits? Entropy.

By anon76203 — On Apr 09, 2010

does it mean that it's dying?

By anon61939 — On Jan 23, 2010

does entropy occur in a human body? i'm confused.


By anon54168 — On Nov 27, 2009

yes, that is exactly right.

By anon16807 — On Aug 15, 2008

If something has high entropy, does it mean that it is more "chaotic"? -Violet

All The Science, in your inbox

Our latest articles, guides, and more, delivered daily.

All The Science, in your inbox

Our latest articles, guides, and more, delivered daily.