Tuesday, June 4, 2019

Entropy And The Second Law Of Thermodynamics

back And The Second Law Of ThermodynamicsThe paper examines, explain clearly, rigorously the term south, then discuss and quantify its meaning in the context of the second law of thermodynamics. Also It will give a historical overview of the term entropy and it will give virtually examples which be taken from the daily life and with these, I will try to explain clearly the term entropy and its intention not only in the context of the second law and also its results in our daily life.2. Introduction (Appendices A.)The term entropy has some related definitions. The first definition applyd by the German physicist Rodolf Julius Clausius in the 1850s and 1860s, he did that to province the second law of thermodynamics. The word entropy has been taken from the Greek word which means transformation. Also just as the first law of thermodynamics leads to the definition of vim as a property of a carcass, so the second law, in the form of Clausius inequality, leads to the definition of a sunrise(prenominal) property of fundamental importance. This property is entropy. In the 1870s the term entropy is given by J. Willard Gibbs. The meaning of what he says is that the entropy shows the uncertainly about the estate of a dodging. The latter can be defining from the probability distribution of its micro-states which demostrates, all molecular details about the system such as the position and the amphetamine of every molecule. If Pi is the possibility of a micro-state i, then the entropy of the system can be expressed byS = -k Pi ln PiWhere k is the Boltzmann constant equal to 1.38062 x10(23) joule/kelvin.Another definition, is the statistical definition developed by Ludwig Boltzmann in 1870s. This definition, describes the entropy as a measure of the image of realizable microscopic configurations of the individual atoms, and molecules of the system which would give rise to the observed macroscopic state of the system.In statistical thermodynamics, Boltzmanns e quation, is a possibility equation relating the Entropy S of an ideal ball up to quantity W, which is the number of micro-states corresponding to a given macro-stateS = k log WWhere k is Boltzmanns equal to 1.38062 x10 (23) joule/kelvin.Boltzmann has proven that the entropy of a given state of thermodynamic al system is connected by a simple relationship to the probability of the state.According to M. Kostic(2004) Entropy is an integral measure of (random) thermal energy redistribution (due to heat transfer or irreversible heat generation) within a system mass and/or piazza (during system expansion), per absolute temperature level. Entropy is increasing from perfectly-ordered (singular and unique) crystalline structure at zero absolute temperature (zero reference) during reversible heating (entropy transfer) and entropy generation during irreversible energy conversion (lost of work-potential to thermal energy), i.e. energy degradation or random equip- partitioning within system ma terial structure and space per absolute temperature level.3. Entropy measures the disorder in a system (Appendices B.)Therefore, metaphorically if a small bookshelf getting disorganized, it will be increasing the entropy of the bookshelf. Because, when the bookshelf is properly organized, finding a book is predictable and easy because all books are in a nice order. As the bookshelf is getting disorganized, the chance of not finding a book increasing, as a result is much higher. So that, when a bookshelf, a room a house are organized and they are moved from creation organized to being disorganized, they generate Entropy. Also, liquids have higher entropy than crystals intuitively because their atomic positions are less orderly. Calculating the entropy of commixture illustrates this interpretation. An example is with scrambling eggs because when we mix the yolk and the white we cannot re-separate after. An example from this situation are given in figures 1.1 and 1,2.V V 2VFig. 1.1 U nmixed atoms. The premixed Fig. 1.2 Mixed atoms. The mixed state N/2state N/2 white atoms on one side, N/2 mixed atoms and N/2 black atoms scatteredblack atoms on the other. with the volume, 2V.Fig. 1.1There are N/2 unnoticeable ideal gas white atoms on one side and N/2 undistinguished gas black atoms on the other side. As a result, the entropy of this systemSunmixed = 2kB logV N/2/(N/2)Twice the configurational entropy of N/2 undistinguished atoms in a volume V. We assume that the black and white atoms have the same masses and the same total energy. Now the entropy change when the partition is removed, as a result from the scrambling and the two sets of atoms allowed mixing. Because, the temperatures and pressures from the both sides are equal and when the partition removing does not involve any heat transfer, and the entropy change to the mixing of the white and black atoms. In desegregated state, the entropy has increased toSmixed = 2kB log(2V )N/2/(N/2)and it isSmixing = Smixe d Sunmixed ==2kB logVN/2/(N/2) / (2V)N/2/(N/2) ==kB log 2N = NkB log 2So that, it gain kB log 2 in entropy every time we place an atom into one of the boxes. James P. Sethna (2006)Furthermore, we can give another(prenominal) example which shows us that entropy measures the disorder in a systemWhich is more disorder?The glass of ice chips or the glass of water?For a glass of water, the number of molecules is astronomical. The ice chips probable look more disorder when we compare to the glass of water which looks uniform. However, according to thermodynamics the ice chips place limits on thenumber of ways the molecules can be arranged. The water molecules in the glass can be arranged in many more ways as a result, they have grater numerosity and therefore greater entropy.4. Entropy measures our ignorance in a systemThe most general is to measure our ignorance about a system. The equilibrium state of a system, maximizes the entropy because, we have lost all information about the initi al conditions, as a result, the entropy maximizing immediately maximises and our ignorance about the details of the system.5. Entropy measures the multiplicity of a systemThe probability of finding a system in a given state depends upon the multiplicity of that state. As a result it is proportional to the number of ways someone can produce that state. Here, it is a pair of dices, and in throwing this pair, that measurable property is the sum of the number of dots which are facing on the top. The multiplicity for two dots showing is just one because there is only one case of the pair that will give that state. For example, the multiplicity for seven dots is six, because there is six cases of the pair that will show a total of seven dots.Probable one way to define the quantity entropy is to do it in terms of the multiplicity.Multiplicity = WEntropy = k lnWWhere K is Boltzmanns constant.For a system, of a large number of particles. We can expect that the system at equilibrium will be f ound in the state of highest multiplicity since the fluctuations from that the state will usually be extremely small to measure. As a result, as a large system approaches equilibrium, its multiplicity therefore, entropy tends obviously to increase. This is one way of stating the Second Law of Thermodynamic.6. The Second Law of Thermodynamics (Appendices C.)The second law of thermodynamics states that heat rises always from the warmer to colder bodies and never opposite. This is a common experience which everyone has seen and probably every day we have a case of those. For example, whenever we pay a cup of warm coffee it will become cool in 10 minutes. The special point of this process is that by the end of long time can never become backward. It has just one direction as time passes. Indeed, through our everyday experience know that when contacting a hot and a cold be will be transferred heat from the hot to the cold body, so the hot body will be a little cold and the cold body the opposite will be a little bit hotter. However, it is never possible as the time passes and the two bodies are in contact the cold body to be colder and the hot body to be hotter, for example, if we put an ice-cube into our drink, the drink does not boil. Therefore, it is only one direction in the flow heat which if we displaced it with a line, then this line will show everything from the past to now and to future.The second law of thermodynamics states that heat cannot be transferred from a colder to a hotter body within a system net changes occurring in other bodies within that system, in any irreversible process, entropy always increases.In nowadays, it is customary to use the term entropy in conjunction with the second law of thermodynamic. Consequently the entropy indicates the unavailable energy of a system, according to the law the entropy of a closed system can never reduce. Another form of the second law thermodynamic says that the minimum amount of heat which exchange a system during a change, which takes place at constant temperature T, associated with a change which is called entropy, with the equationdQ=

No comments:

Post a Comment

Note: Only a member of this blog may post a comment.