**Author**: Claude E Shannon,Warren Weaver

**Publisher:**University of Illinois Press

**ISBN:**025209803X

**Category:**Language Arts & Disciplines

**Page:**144

**View:**6104

Skip to content
# Search Results for: the-mathematical-theory-of-communication

**Author**: Claude E Shannon,Warren Weaver

**Publisher:** University of Illinois Press

**ISBN:** 025209803X

**Category:** Language Arts & Disciplines

**Page:** 144

**View:** 6104

Scientific knowledge grows at a phenomenal pace--but few books have had as lasting an impact or played as important a role in our modern world as The Mathematical Theory of Communication, published originally as a paper on communication theory more than fifty years ago. Republished in book form shortly thereafter, it has since gone through four hardcover and sixteen paperback printings. It is a revolutionary work, astounding in its foresight and contemporaneity. The University of Illinois Press is pleased and honored to issue this commemorative reprinting of a classic.

**Author**: Jan Kåhre

**Publisher:** Springer Science & Business Media

**ISBN:** 1461509750

**Category:** Technology & Engineering

**Page:** 502

**View:** 2747

The general concept of information is here, for the first time, defined mathematically by adding one single axiom to the probability theory. This Mathematical Theory of Information is explored in fourteen chapters: 1. Information can be measured in different units, in anything from bits to dollars. We will here argue that any measure is acceptable if it does not violate the Law of Diminishing Information. This law is supported by two independent arguments: one derived from the Bar-Hillel ideal receiver, the other is based on Shannon's noisy channel. The entropy in the 'classical information theory' is one of the measures conforming to the Law of Diminishing Information, but it has, however, properties such as being symmetric, which makes it unsuitable for some applications. The measure reliability is found to be a universal information measure. 2. For discrete and finite signals, the Law of Diminishing Information is defined mathematically, using probability theory and matrix algebra. 3. The Law of Diminishing Information is used as an axiom to derive essential properties of information. Byron's law: there is more information in a lie than in gibberish. Preservation: no information is lost in a reversible channel. Etc. The Mathematical Theory of Information supports colligation, i. e. the property to bind facts together making 'two plus two greater than four'. Colligation is a must when the information carries knowledge, or is a base for decisions. In such cases, reliability is always a useful information measure. Entropy does not allow colligation.
*The Theory of Probability*

**Author**: Warren Weaver

**Publisher:** Courier Corporation

**ISBN:** 0486150917

**Category:** Mathematics

**Page:** 400

**View:** 9215

This witty, nontechnical introduction to probability elucidates such concepts as permutations, independent events, mathematical expectation, the law of averages and more. No advanced math required. 49 drawings.
*The Self-organization of the Knowledge-based Society*

**Author**: Loet Leydesdorff

**Publisher:** Universal-Publishers

**ISBN:** 1581126956

**Category:** Social Science

**Page:** 351

**View:** 2285

Networks of communication evolve in terms of reflexive exchanges. The codification of these reflections in language, that is, at the social level, can be considered as the operating system of society. Under sociologically specifiable conditions, the discursive reconstructions can be expected to make the systems under reflection increasingly knowledge-intensive. This sociological theory of communication is founded in a tradition that includes Giddens' (1979) structuration theory, Habermas' (1981) theory of communicative action, and Luhmann's (1984) proposal to consider social systems as self-organizing. The study also elaborates on Shannon's (1948) mathematical theory of communication for the formalization and operationalization of the non-linear dynamics. The development of scientific communications can be studied using citation analysis. The exchange media at the interfaces of knowledge production provide us with the evolutionary model of a Triple Helix of university-industry-government relations. The construction of the European Information Society can then be analyzed in terms of interacting networks of communication. The issues of sustainable development and the expectation of social change are discussed in relation to the possibility of a general theory of communication. REVIEW In this book, LoetLeydesdorff sets out to answer the question, "Can society be considered as a self-organizing (autopoietic) system. In the process, Leydesdorff, develops a general sociological theory of communication, as well as a special theory of scientific communication designed to analyze complex systems such as the Euroean Information Society. (from review in JASIST 53[1], 2002, 62-63)
*How Claude Shannon Invented the Information Age*

**Author**: Jimmy Soni,Rob Goodman

**Publisher:** Simon and Schuster

**ISBN:** 1476766703

**Category:** Biography & Autobiography

**Page:** 384

**View:** 750

“A charming account of one of the twentieth century’s most distinguished scientists…a modern-day Da Vinci” (Fortune) whose insights stand behind every computer built, email sent, video streamed, and webpage loaded—“without Shannon, the digital revolution would have ground to a halt” (The Wall Street Journal). Claude Shannon was a groundbreaking polymath, a brilliant tinkerer, and a digital pioneer. He constructed a fleet of customized unicycles and a flamethrowing trumpet, outfoxed Vegas casinos, and built juggling robots. He also wrote the seminal text of the digital revolution, which has been called “the Magna Carta of the Information Age.” His discoveries would lead contemporaries to compare him to Albert Einstein and Isaac Newton. His work anticipated by decades the world we’d be living in today—and gave mathematicians and engineers the tools to bring that world to pass. In this “welcome and inspiring account of a largely unsung hero” (Kirkus Reviews), Jimmy Soni and Rob Goodman reveal Claude Shannon’s full story for the first time. It’s the story of a small-town Michigan boy whose career stretched from the era of room-sized computers powered by gears and string to the age of Apple. It’s the story of the origins of our digital world in the tunnels of MIT and the “idea factory” of Bell Labs, in the “scientists’ war” with Nazi Germany, and in the work of Shannon’s collaborators and rivals, thinkers like Alan Turing, John von Neumann, Vannevar Bush, and Norbert Wiener. And it’s the story of Shannon’s life as an often reclusive, always playful genius. “We owe Claude Shannon a lot, and Soni and Goodman’s book takes a big first step in paying that debt” (San Francisco Review of Books). With access to Shannon’s family and friends, A Mind at Play is “a boon for those eager to know more about his incredibly influential life—whimsical, independent and curiosity-driven...a vivid portrayal” (Nature).
*collected papers*

**Author**: Claude Elwood Shannon,Neil James Alexander Sloane,Aaron D. Wyner,IEEE Information Theory Society

**Publisher:** Wiley-IEEE Press

**ISBN:** 9780780304345

**Category:** Biography & Autobiography

**Page:** 924

**View:** 8454

This important book, the first published collection of papers by Claude E. Shannon, is a fascinating guide to all of the published articles from this world-renowned inventor, tinkerer, puzzle-solver, prankster, and father of information theory. Includes his seminal article THE MATHEMATICAL THEORY OF COMMUNICATION.

**Author**: Glenn Shafer

**Publisher:** Princeton University Press

**ISBN:** 069110042X

**Category:** Business & Economics

**Page:** 297

**View:** 7696

Both in science and in practical affairs we reason by combining facts only inconclusively supported by evidence. Building on an abstract understanding of this process of combination, this book constructs a new theory of epistemic probability. The theory draws on the work of A. P. Dempster but diverges from Depster's viewpoint by identifying his "lower probabilities" as epistemic probabilities and taking his rule for combining "upper and lower probabilities" as fundamental. The book opens with a critique of the well-known Bayesian theory of epistemic probability. It then proceeds to develop an alternative to the additive set functions and the rule of conditioning of the Bayesian theory: set functions that need only be what Choquet called "monotone of order of infinity." and Dempster's rule for combining such set functions. This rule, together with the idea of "weights of evidence," leads to both an extensive new theory and a better understanding of the Bayesian theory. The book concludes with a brief treatment of statistical inference and a discussion of the limitations of epistemic probability. Appendices contain mathematical proofs, which are relatively elementary and seldom depend on mathematics more advanced that the binomial theorem.
*50 Years of Discovery*

**Author**: Sergio Verdú,Steven W. McLaughlin,IEEE Information Theory Society

**Publisher:** Wiley-IEEE Press

**ISBN:** 9780780353633

**Category:** Technology & Engineering

**Page:** 758

**View:** 1987

Celebrating 50 years since the discovery of information theory by Claude Shannon, this book consists to the 50 best tutorials in the area compiled by the editors of the "IEEE Transactions on Information Theory." These articles cover the technologies at the heart of communications, signal processing, computer and control systems and serve as a valuable guide for all those interested in the basis for information theory.

**Author**: Aleksandr I?Akovlevich Khinchin

**Publisher:** Courier Corporation

**ISBN:** 0486604349

**Category:** Mathematics

**Page:** 120

**View:** 8874

First comprehensive introduction to information theory explores the work of Shannon, McMillan, Feinstein, and Khinchin. Topics include the entropy concept in probability theory, fundamental theorems, and other subjects. 1957 edition.
*Second Edition*

**Author**: Frank K Hwang

**Publisher:** World Scientific Publishing Company

**ISBN:** 9813106425

**Category:**

**Page:** 200

**View:** 5605

The first edition of this book covered in depth the mathematical theory of nonblocking multistage interconnecting networks, which is applicable to both communication and computer networks. This comprehensively updated version puts more emphasis to the multicast and multirate networks which are under fast development recently due to their wide applications. This comprehensively updated new edition not only introduces the classical theory of the fundamental point-to-point network but also has a renewed emphasis on the latest multicast and multirate networks. The book can serve as either a one- or two-semester textbook for graduate students of information science, (electronic) communications, and applied mathematics. In addition, as all the relevant literature is organized and evaluated under one structured framework, the volume is an essential reference for researchers in those areas.
*A Statistical Theory of Communications*

**Author**: Robert M. Fano

**Publisher:** N.A

**ISBN:** N.A

**Category:** Information theory

**Page:** 389

**View:** 9808

**Author**: Robert M. Gray

**Publisher:** Springer Science & Business Media

**ISBN:** 1475739826

**Category:** Computers

**Page:** 332

**View:** 7566

*Body, Flesh, and Relationship*

**Author**: Frank J. Macke

**Publisher:** Rowman & Littlefield

**ISBN:** 161147549X

**Category:** Language Arts & Disciplines

**Page:** 258

**View:** 648

The Experience of Human Communication approaches everyday communication as a philosophical and psychological matter. Using insights from Merleau-Ponty, Heidegger, and Foucault, Frank Macke stresses that human communication—and with it, the human body—is, first and foremost, a relational phenomenon involving friends and family.
*Symbols, Signals and Noise*

**Author**: John R. Pierce

**Publisher:** Courier Corporation

**ISBN:** 0486134970

**Category:** Computers

**Page:** 336

**View:** 3262

Covers encoding and binary digits, entropy, language and meaning, efficient encoding and the noisy channel, and explores ways in which information theory relates to physics, cybernetics, psychology, and art. 1980 edition.

**Author**: Robert B. Ash

**Publisher:** Courier Corporation

**ISBN:** 0486141454

**Category:** Technology & Engineering

**Page:** 352

**View:** 688

DIVAnalysis of channel models and proof of coding theorems; study of specific coding systems; and study of statistical properties of information sources. Sixty problems, with solutions. Advanced undergraduate to graduate level. /div
*A History, a Theory, a Flood*

**Author**: James Gleick

**Publisher:** N.A

**ISBN:** 9780007225743

**Category:** Communication

**Page:** 526

**View:** 2059

From the invention of scripts and alphabets to the long misunderstood "talking drums" of Africa, James Gleick tells the story of information technologies that changed the very nature of human consciousness. He also provides portraits of the key figures contributing to the inexorable development of our modern understanding of information, including Charles Babbage, Ada Byron, Samuel Morse, Alan Turing, and Claude Shannon.
*A Tutorial Introduction*

**Author**: JV Stone

**Publisher:** Sebtel Press

**ISBN:** 0956372856

**Category:** Information theory

**Page:** 243

**View:** 6812

Originally developed by Claude Shannon in the 1940s, information theory laid the foundations for the digital revolution, and is now an essential tool in telecommunications, genetics, linguistics, brain sciences, and deep space communication. In this richly illustrated book, accessible examples are used to introduce information theory in terms of everyday games like ‘20 questions’ before more advanced topics are explored. Online MatLab and Python computer programs provide hands-on experience of information theory in action, and PowerPoint slides give support for teaching. Written in an informal style, with a comprehensive glossary and tutorial appendices, this text is an ideal primer for novices who wish to learn the essential principles and applications of information theory.

Full PDF Download Free

Privacy Policy

Copyright © 2018 Download PDF Site — Primer WordPress theme by GoDaddy