Hegemony refers to the dominance of one group over others, not just through force but by shaping cultural norms and values to make its power seem natural and legitimate. Power, in this context, is exercised not only through direct control but also through the consent of the governed, achieved by embedding the dominant group's worldview into the fabric of society.