• Wyszukiwanie zaawansowane
  • Kategorie
  • Kategorie BISAC
  • Książki na zamówienie
  • Promocje
  • Granty
  • Książka na prezent
  • Opinie
  • Pomoc
  • Załóż konto
  • Zaloguj się

Information Theory: Three Theorems by Claude Shannon » książka

zaloguj się | załóż konto
Logo Krainaksiazek.pl

koszyk

konto

szukaj
topmenu
Księgarnia internetowa
Szukaj
Książki na zamówienie
Promocje
Granty
Książka na prezent
Moje konto
Pomoc
 
 
Wyszukiwanie zaawansowane
Pusty koszyk
Bezpłatna dostawa dla zamówień powyżej 20 złBezpłatna dostawa dla zamówień powyżej 20 zł

Kategorie główne

• Nauka
 [2946600]
• Literatura piękna
 [1856966]

  więcej...
• Turystyka
 [72221]
• Informatyka
 [151456]
• Komiksy
 [35826]
• Encyklopedie
 [23190]
• Dziecięca
 [619653]
• Hobby
 [140543]
• AudioBooki
 [1577]
• Literatura faktu
 [228355]
• Muzyka CD
 [410]
• Słowniki
 [2874]
• Inne
 [445822]
• Kalendarze
 [1744]
• Podręczniki
 [167141]
• Poradniki
 [482898]
• Religia
 [510455]
• Czasopisma
 [526]
• Sport
 [61590]
• Sztuka
 [243598]
• CD, DVD, Video
 [3423]
• Technologie
 [219201]
• Zdrowie
 [101638]
• Książkowe Klimaty
 [124]
• Zabawki
 [2473]
• Puzzle, gry
 [3898]
• Literatura w języku ukraińskim
 [254]
• Art. papiernicze i szkolne
 [8170]
Kategorie szczegółowe BISAC

Information Theory: Three Theorems by Claude Shannon

ISBN-13: 9783031215605 / Angielski / Miękka / 2023

Antoine Chambert-Loir
Information Theory: Three Theorems by Claude Shannon Antoine Chambert-Loir 9783031215605 Springer - książkaWidoczna okładka, to zdjęcie poglądowe, a rzeczywista szata graficzna może różnić się od prezentowanej.

Information Theory: Three Theorems by Claude Shannon

ISBN-13: 9783031215605 / Angielski / Miękka / 2023

Antoine Chambert-Loir
cena 262,25 zł
(netto: 249,76 VAT:  5%)

Najniższa cena z 30 dni: 250,57 zł
Termin realizacji zamówienia:
ok. 22 dni roboczych
Bez gwarancji dostawy przed świętami

Darmowa dostawa!

This book provides an introduction to information theory, focussing on Shannon’s three foundational theorems of 1948–1949. Shannon’s first two theorems, based on the notion of entropy in probability theory, specify the extent to which a message can be compressed for fast transmission and how to erase errors associated with poor transmission. The third theorem, using Fourier theory, ensures that a signal can be reconstructed from a sufficiently fine sampling of it. These three theorems constitute the roadmap of the book.The first chapter studies the entropyof a discrete random variable and related notions. The second chapter, on compression and error correcting, introduces the concept of coding, proves the existence of optimal codes and good codes (Shannon's first theorem), and shows how information can be transmitted in the presence of noise (Shannon's second theorem). The third chapter proves the sampling theorem (Shannon's third theorem) and looks at its connections with other results, such as the Poisson summation formula. Finally, there is a discussion of the uncertainty principle in information theory.Featuring a good supply of exercises (with solutions), and an introductory chapter covering the prerequisites, this text stems out lectures given to mathematics/computer science students at the beginning graduate level.

This book provides an introduction to information theory, focussing on Shannon’s three foundational theorems of 1948–1949. Shannon’s first two theorems, based on the notion of entropy in probability theory, specify the extent to which a message can be compressed for fast transmission and how to erase errors associated with poor transmission. The third theorem, using Fourier theory, ensures that a signal can be reconstructed from a sufficiently fine sampling of it. These three theorems constitute the roadmap of the book. The first chapter studies the entropy of a discrete random variable and related notions. The second chapter, on compression and error correcting, introduces the concept of coding, proves the existence of optimal codes and good codes (Shannon's first theorem), and shows how information can be transmitted in the presence of noise (Shannon's second theorem). The third chapter proves the sampling theorem (Shannon's third theorem) and looks at its connections with other results, such as the Poisson summation formula. Finally, there is a discussion of the uncertainty principle in information theory.Featuring a good supply of exercises (with solutions), and an introductory chapter covering the prerequisites, this text stems out lectures given to mathematics/computer science students at the beginning graduate level.

Kategorie:
Informatyka, Bazy danych
Kategorie BISAC:
Computers > Data Science - General
Computers > Information Theory
Computers > Computer Science
Wydawca:
Springer
Język:
Angielski
ISBN-13:
9783031215605
Rok wydania:
2023
Dostępne języki:
Oprawa:
Miękka

Elements of Theory of Probability.- Entropy and Mutual Information.- Coding.- Sampling.- Solutions to Exercises.- Bibliography.- Notation.- Index.


Antoine Chambert-Loir is a professor of mathematics at Université Paris Cité. His research addresses questions in algebraic geometry which are motivated by number theoretical problems. He is the author of two books published by Springer-Verlag: A Field Guide To Algebra, an introduction to Galois theory; and (Mostly) Commutative Algebra, an intermediate-level exposition of commutative algebra. With  J. Nicaise and J. Sebag, he cowrote the research monograph Motivic Integration (published by Birkhäuser), which was awarded the 2017 Ferran Sunyer i Balaguer prize.


This book provides an introduction to information theory, focussing on Shannon’s three foundational theorems of 1948–1949. Shannon’s first two theorems, based on the notion of entropy in probability theory, specify the extent to which a message can be compressed for fast transmission and how to erase errors associated with poor transmission. The third theorem, using Fourier theory, ensures that a signal can be reconstructed from a sufficiently fine sampling of it. These three theorems constitute the roadmap of the book. 

The first chapter studies the entropy of a discrete random variable and related notions. The second chapter, on compression and error correcting, introduces the concept of coding, proves the existence of optimal codes and good codes (Shannon's first theorem), and shows how information can be transmitted in the presence of noise (Shannon's second theorem). The third chapter proves the sampling theorem (Shannon's third theorem) and looks at its connections with other results, such as the Poisson summation formula. Finally, there is a discussion of the uncertainty principle in information theory.

Featuring a good supply of exercises (with solutions), and an introductory chapter covering the prerequisites, this text stems out lectures given to mathematics/computer science students at the beginning graduate level.


Chambert-Loir, Antoine Antoine Chambert-Loir is Professor at Universit?? ... więcej >


Udostępnij

Facebook - konto krainaksiazek.pl



Opinie o Krainaksiazek.pl na Opineo.pl

Partner Mybenefit

Krainaksiazek.pl w programie rzetelna firma Krainaksiaze.pl - płatności przez paypal

Czytaj nas na:

Facebook - krainaksiazek.pl
  • książki na zamówienie
  • granty
  • książka na prezent
  • kontakt
  • pomoc
  • opinie
  • regulamin
  • polityka prywatności

Zobacz:

  • Księgarnia czeska

  • Wydawnictwo Książkowe Klimaty

1997-2025 DolnySlask.com Agencja Internetowa

© 1997-2022 krainaksiazek.pl
     
KONTAKT | REGULAMIN | POLITYKA PRYWATNOŚCI | USTAWIENIA PRYWATNOŚCI
Zobacz: Księgarnia Czeska | Wydawnictwo Książkowe Klimaty | Mapa strony | Lista autorów
KrainaKsiazek.PL - Księgarnia Internetowa
Polityka prywatnosci - link
Krainaksiazek.pl - płatnośc Przelewy24
Przechowalnia Przechowalnia