BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//Talks.cam//talks.cam.ac.uk//
X-WR-CALNAME:Talks.cam
BEGIN:VEVENT
SUMMARY:Understanding the Source Coding Theorem: A Talk on Shannon’s Ent
 ropy - Kwing Hei Li\, Churchill College
DTSTART:20201021T180000Z
DTEND:20201021T183000Z
UID:TALK153460@talks.cam.ac.uk
CONTACT:Matthew Ireland
DESCRIPTION:How many bits do we need to encode a sequence of English chara
 cters? Can we do better by considering the relative frequency of each char
 acter? Is there a theoretical limit to how we encode data with negligible 
 risk of information loss?\n\nIn this talk\, we first define Shannon’s en
 tropy\, which quantifies the predictability of a sequence of random variab
 les. We will also explore the Source Coding Theorem\, which provides an op
 erational definition of entropy by establishing a fundamental limit to the
  compressibility of information. Finally\, we will prove this theorem with
  some surprisingly simple results in Probability Theory.
LOCATION:Online\, via MS Teams
END:VEVENT
END:VCALENDAR
