Presentation on theme: "Introduction to Multimedia1 Introduction to Broadband Multimedia Network."— Presentation transcript:
Introduction to Multimedia1 Introduction to Broadband Multimedia Network
Introduction to Multimedia2 Introduction zScope of Broadband zMultimedia Description zWhy multimedia systems? zClassification of Media zMultimedia Systems zData Stream Characteristics
BROADBAND Broadband Signifies : High Bandwidth High Access speeds, 256 Kbps t o 100 Mbps Huge Core bandwidth pipes, STM –16 (SDH), GigE (MEN) and 2.5 GigE (DWDM / CWDM) Multiple Converged Services High Speed Data Voice Video
07/26/04Page - 4 Multiple Definitions zBroadband The capability of supporting, in both the provider-to-consumer (downstream) and the consumer-to-provider (upstream) directions, a speed in excess of 200 kilobits per second (kbps) in the last mile FCC 1999 Telecommunications Act Deployment Report z“High-speed” Services with over 200 kbps capability in at least one direction. The term high- speed services includes advanced telecommunications capability zThe International Telecommunications Union’s (ITU) defines broadband service as 1.5 Mbps
07/26/04Page - 5 Speed Equals Time Downloading the DVD Movie “The Matrix” 7.8 GB
07/26/04Page - 7 Technology Futures, Inc. (2001) zThe typical household of 2015 subscribes to broadband service at 24 Mb/s to100 Mb/s, zSmall businesses will access the network at data rates up to 622 Mb/s. zMedium and large businesses will access the network directly with fiber at data rates from 2.4 Gb/s to 40 Gb/s. zBy 2015, most customers obtain voice and narrowband data service via wireless or VoIP on broadband channels. zIn 2015, fiber dominates the outside plant, comprising 100% of the interoffice network, 97% of the feeder network, and 95% of the distribution network.
BROADBAND SERVICES Services Offered On Broadband : Data Services High Speed Internet Services Point to Point and Point to Multi Point VPN Services Web Hosting Applications Walled garden (Internet on TV) Voices Services Audio Conference Voice over IP (VoIP ) Video Services Video Broadcast Video on Demand Video telephony Online Gaming
BROADBAND SERVICES Basic Telecom Service Voice Internet Value-added Applications Time Web Hosting Multimedia Conference Video/Audio- On-Demand Online Gaming Video on PC Revenue Value Add applications to boost revenues Services Offered on Broadband
BROADBAND COMPONENTS video communication gaming video IP VPN Business Residential Voice over IP Core Network Access & Aggregation Network NMS Server Radius Server NMS Client Video Server User Backend CoreAccess Soft Switch
BROADBAND NETWORK ELEMENTS The Access The Core Wired Access ( DSL, Cable, FTTH) Wireless ( BWA, Wifi ) The Media Optical Fiber The Technology SDH CWDM / DWDM ATM Ethernet Backend Authentication Server Content Servers Web Hosting Servers Video Servers Network Elements – Technology Options Corporate / SME / SOHO Residential User
BROADBAND NETWORK COMPONENTS Dual Homed Ring 2 Fiber SM G.652 DSLAM Internet MPLS BBRAS Internet Gateway Router GigE Backend DNS DHCP AAA Core Network Access Voice Services Server Internet Data Services Server TV/Video Services Servers FTTH 100M M/C STB
CORE NETWORK TECHNOLOGIES Broadband Core Technologies : Core Network : Architecture Ring Single Homing Dual Homing Mesh Technology ATM / IP / Ethernet SDH SDH over CWDM / DWDM
RING ARCHITECTURE Aggregation Ring 2 Fiber SM G.652 Core Switch Collector Ring Advantage Low Cost Simple Architecture Disadvantage Hub – Potential Single Point failure Target Application Towards Access and Low Capacity Core Advantage Highly Reliable Disadvantage High Cost Solution. Target Application High Capacity Core Single Homed Dual Homed
MESH ARCHITECTURE Core Ring Dual-Homed Mesh Advantages Extremely Reliable Protection against Equipment / Fiber failure Disadvantages Very Complex and Costly to implement Target Application Highest Tier in Core Mesh Network
CORE TECHNOLOGIES Technology Options on Core 1.ATM Backbone 2.IP / Metro Ethernet 3.SDH / CWDM / DWDM.
ATM BACKBONE Traditional ATM Backbone Core Benefits: 1.High deployment across the world 2.Stabilized Technology 3.Aggregation through multiple E1’s or upto STM 16 rings. 4.Good for aggregating Higher bandwidths. Drawbacks: 1.ATM pvc uses 10% overheads 2.Higher Provisioning time.
ATM BACKBONE CORE E D Broadband RAS A STM4/STM16 DSLAM for Residential Internet Access– Traditional Way xDSL B ATM MESH RFC 1483 Bridged Access RFC 1483 Bridged Access Internet RFC 1483 Bridged Access STM4/STM16 FTTH 100M LAN Switch Metro Ethernet
METRO ETHERNET CORE Metro Ethernet Core Benefits: 1.Deployment has started in Huge way across the world 2.Aggregation through Fast Ethernet or Gigabit Ethernet. 3.Highly recommended for high bandwidths requirements. Drawbacks: 1.Technology getting standardized 2.Limited support for VLAN’s. 3.Single broadcast domain.
BROADBAND TECHNOLOGIES & SERVICES Ethernet Service Connectivity Provider DSLAM for Residential Internet Access– Next Generation FE/Gig xDSL Internet Gateway Router Broadband RAS IPDSLAM
SDH / CWDM / DWDM CORE SDH Network STM 1/4/16 Aggregation Ring (CWDM/DWDM) 2 Fiber SM G.652 DSLAM traffic going to the core through Metro Ethernet xDSL Metro Transport and Aggregation Network Inter-City Backbone Network Ethernet UNI Internet MPLS BBRAS Internet Access Router Internet Gateway Router GigE Backbone SDH STM1/STM4 on Sonet ATM STM1/STM4 on ATM Core Switch Voice Services Server Internet Data Services Server TV/Video Services Servers DNS DHCP AAA Aggregation Node
Network Topologies zA topology refers to the manner in which the cable is run to individual workstations on the network. ythe configurations formed by the connections between devices on a local area network (LAN) or between two or more LANs zThere are three basic network topologies (not counting variations thereon): the bus, the star, and the ring. zIt is important to make a distinction between a topology and an architecture. yA topology is concerned with the physical arrangement of the network components. yIn contrast, an architecture addresses the components themselves and how a system is structured (cable access methods, lower level protocols, topology, etc.). An example of an architecture is 10baseT Ethernet which typically uses the star topology.
Bus Topology zA bus topology connects each computer (node) to a single segment trunk. yA ‘trunk’ is a communication line, typically coax cable, that is referred to as the ‘bus.’ The signal travels from one end of the bus to the other. yA terminator is required at each end to absorb the signal so it does not reflect back across the bus. zIn a bus topology, signals are broadcast to all stations. Each computer checks the address on the signal (data frame) as it passes along the bus. If the signal’s address matches that of the computer, the computer processes the signal. If the address doesn’t match, the computer takes no action and the signal travels on down the bus. zOnly one computer can ‘talk’ on a network at a time. A media access method (protocol) called CSMA/CD is used to handle the collisions that occur when two signals are placed on the wire at the same time. zThe bus topology is passive. In other words, the computers on the bus simply ‘listen’ for a signal; they are not responsible for moving the signal along. zA bus topology is normally implemented with coaxial cable.
Bus Topology zAdvantages of bus topology: yEasy to implement and extend yWell suited for temporary networks that must be set up in a hurry yTypically the cheapest topology to implement yFailure of one station does not affect others zDisadvantages of bus topology: yDifficult to administer/troubleshoot yLimited cable length and number of stations yA cable break can disable the entire network; no redundancy yMaintenance costs may be higher in the long run yPerformance degrades as additional computers are added
Star Topology zAll of the stations in a star topology are connected to a central unit called a hub. yThe hub offers a common connection for all stations on the network. Each station has its own direct cable connection to the hub. In most cases, this means more cable is required than for a bus topology. However, this makes adding or moving computers a relatively easy task; simply plug them into a cable outlet on the wall. zIf a cable is cut, it only affects the computer that was attached to it. This eliminates the single point of failure problem associated with the bus topology. (Unless, of course, the hub itself goes down.) zStar topologies are normally implemented using twisted pair cable, specifically unshielded twisted pair (UTP). The star topology is probably the most common form of network topology currently in use.
Star Topology zAdvantages of star topology: yEasy to add new stations yEasy to monitor and troubleshoot yCan accommodate different wiring zDisadvantages of star topology: yFailure of hub cripples attached stations yMore cable required (more expensive to wire a building for networking)
Ring Topology zA ring topology consists of a set of stations connected serially by cable. In other words, it’s a circle or ring of computers. There are no terminated ends to the cable; the signal travels around the circle in a clockwise (or anticlockwise) direction. zNote that while this topology functions logically as ring, it is physically wired as a star. The central connector is not called a hub but a Multistation Access Unit or MAU. (Don’t confuse a Token Ring MAU with a ‘Media Adapter Unit’ which is actually a transceiver.) zUnder the ring concept, a signal is transferred sequentially via a "token" from one station to the next. When a station wants to transmit, it "grabs" the token, attaches data and an address to it, and then sends it around the ring. The token travels along the ring until it reaches the destination address. The receiving computer acknowledges receipt with a return message to the sender. The sender then releases the token for use by another computer. zEach station on the ring has equal access but only one station can talk at a time.
Ring Topology zIn contrast to the ‘passive’ topology of the bus, the ring employs an ‘active’ topology. Each station repeats or ’boosts’ the signal before passing it on to the next station. zRings are normally implemented using twisted pair or fiber-optic cable zAdvantages of ring topology: yGrowth of system has minimal impact on performance yAll stations have equal access zDisadvantages of ring topology: yMost expensive topology yFailure of one computer may impact others yComplex
What’s “New Generation Network” or NWGN? Present Network Revised NXGN New Generation Network (NWGN) Past Network Next Generation Network (NXGN) 2) modification 1) clean-slate Examples: Cell Phones > 2G > 3G > 4G? Internet > IPv4 > IPv6 > IPv? Next Generations New Generations
Introduction to Multimedia30 Broadband in Indonesia
sps From Agricultural to Conceptual
32 The Information Revolution, Driver of the Knowledge Economy in a Global World
33 ROLE OF BROADBAND "for every one percentage point increase in broadband penetration in a state, employment is projected to increase by 0.2 to 0.3 percent per year” (brooking institute)
34 ROLE OF BROADBAND Broadband needs to be considered as basic national infrastructure, as it will fundamentally reshape the world in the 21st century and change the way services are delivered – from e-health to e-education to e-commerce to e- government. Broadband is the most powerful tool ever devised to drive social and economic development, and accelerate progress towards the Millennium Development Goals. Broadband is becoming a prerequisite to economic opportunity for individuals, small businesses and communities. Those without broadband and the skills to use broadband-enabled technologies are becoming more isolated from the modern American economy. Broadband can provide significant benefits to the next generation of entrepreneurs and small businesses—the engines of job creation and economic growth for the country.
35 BROADBAND & SMEs It allows small businesses to achieve operational scale more quickly. Broadband and associated ICTs can help lower company start-up costs through faster business registration and improved access to customers and suppliers. It gives SMEs access to new markets and opportunities by lowering the barriers of physical scale and allowing them to compete for customers who previously turned exclusively to larger suppliers. It allow small businesses to increase efficiency, improve market access, reduce costs and increase the speed of both transactions and interactions. E-commerce solutions eliminate geographic barriers to getting a business's message and product out to a broad audience. 60 million Americans go online every day to find a product or service, but only 24% of small businesses use e-commerce applications to sell online
36 BROADBAND & ECONOMIC SECTORS zOECD report urges governments to invest in open-access high-speed national fiber networks that can serve as the future delivery mechanism for a huge range of new and innovative public sector services. zAnd despite the large initial capital investment needed – typically US$ 1,500- 2,500 per household connected – the report shows that National Broadband Networks can pay for themselves within ten years, through dramatic savings in just four key economic sectors: zelectricity zhealthcare zroad transportroad transport zEducationEducation zcost savings across the four sectors of just 0.5%-1.5% would be sufficient to justify the cost of laying high-speed fiber-to-the- home via a national point-to-point network.
37 The Positive Side of Indonesian ICT Development Mobile and Internet Tariffs are among the cheapest in SE Asia Large growths in Mobile Subscribers for several years The growing applications and contents in Internet and Mobile services, such as IP-TV, streaming videos, games, entertainments, BlackBerry, etc. Indonesia is among the World's largest users of Web 2.0 Social Networking, such as Blogs, Facebook, Multiply, Youtube, YM, Chatting, etc
38 The Negative Side of the Indonesian ICT Development zThe declining profit margins of Operators due to very intense tariff competition zThe lowering of Quality of Service, especially 3G and mobile Internet services zLow or little profits from Web, Internet and Social Networks, due to average low income of Indonesians zICT growth has not been accompanied by economic growth; little value added results
39 ICT Indicators Population in 2008 = 228,523,300 Households in 2008 = 57,716,100 Income per Capita = Rp 7.5 millions PDB per Capita = Rp 8.7 millions per year % of Households with Fixed Phones = 12.69% (24.51% in cities, 3.72% in villages)
40 INFRASTRUKTUR DATA 2008 INDONESIA (UN E-Gov Survey 2008) Internet / 100 Users7.18 PC / 100 Users1.47 Cellular Subs /100 users28.30 Main Telephone Lines/100 Users6.57 Broadband / 100 Users0.05
41 Negara Peringkat Kesiapan Teknologi (Sumber: Global Competitiveness Report , World Economic Forum) Daya Saing Daya Saing Teknologi Tekno- logi Maju Daya Serap Teknologi Regulasi TIK FDI dan Transfer Teknologi Jasa Seluler Pengguna Internet Jumlah Komputer Broad- band Thailand Indonesia Vietnam Philipina Sri Lanka Kamboja e- Readiness 2008 (Sumber: The Economist Intelligence Unit, 2007) NegaraPeringkatNilai TotalAksesBisnisSos BudHukumKebijakan Adopsi Bisnis Thailand475,223,806,995,075,905,255,10 Philipina554,903,206,564,534,505,205,45 Sri Lanka604,352,955,804,806,304,103,70 Vietnam654,032,256,313,804,404,603,75 Indonesia683,592,306,493,533,203,403,20 E-Readiness Sumber : RPJMN
42 WHY FIXED BROADBAND ? Mostly dedicated sampai ke last-miles Wireless pada umumnya untuk low-traffic Infrastruktur Dasar Long-term investment Public Private Partnership Optimalisasi Pemanfaatan Palapa Ring Industri Kreatif sangat membutuhkan
Introduction to Multimedia43 Multimedia Description zMultimedia xis an integration of continuous media (e.g. audio, video) and discrete media (e.g. text, graphics, images) through which digital information can be conveyed to the user in an appropriate way. yMulti xmany, much, multiple yMedium xAn interleaving substance through which something is transmitted or carried on
Introduction to Multimedia44 Why Multimedia Computing? yApplication driven xe.g. medicine, sports, entertainment, education yInformation can often be better represented using audio/video/animation rather than using text, images and graphics alone. yInformation is distributed using computer and telecommunication networks. yIntegration of multiple media places demands on xcomputation power xstorage requirements xnetworking requirements
Introduction to Multimedia45 Multimedia Information Systems zTechnical challenges ySheer volume of data xNeed to manage huge volumes of data yTiming requirements xamong components of data computation and communication. xMust work internally with given timing constraints - real-time performance is required. yIntegration requirements xneed to process traditional media (text, images) as well as continuous media (audio/video). xMedia are not always independent of each other - synchronization among the media may be required.
Introduction to Multimedia46 High Data Volume of Multimedia Information
Introduction to Multimedia47 Technology Incentive zGrowth in computational capacity xMM workstations with audio/video processing capability xDramatic increase in CPU processing power xDedicated compression engines for audio, video etc. zRise in storage capacity xLarge capacity disks (several gigabytes) xIncrease in storage bandwidth,e.g. disk array technology zSurge in available network bandwidth xhigh speed fiber optic networks - gigabit networks xfast packet switching technology
Introduction to Multimedia48 Application Areas zResidential Services xvideo-on-demand xvideo phone/conferencing systems xmultimedia home shopping (MM catalogs, product demos and presentation) xself-paced education zBusiness Services xCorporate training xDesktop MM conferencing, MM
Introduction to Multimedia49 Application Areas zEducation xDistance education - MM repository of class videos xAccess to digital MM libraries over high speed networks zScience and Technology xcomputational visualization and prototyping xastronomy, environmental science zMedicine xDiagnosis and treatment - e.g. MM databases that provide support for queries on scanned images, X-rays, assessments, response etc.
Introduction to Multimedia50 Classification of Media yPerception Medium xHow do humans perceive information in a computer? Through seeing - text, images, video Through hearing - music, noise, speech yRepresentation Medium xHow is the computer information encoded? Using formats for representing and information ASCII(text), JPEG(image), MPEG(video) yPresentation Medium xThrough which medium is information delivered by the computer or introduced into the computer? Via I/O tools and devices paper, screen, speakers (output media) keyboard, mouse, camera, microphone (input media)
Introduction to Multimedia51 Classification of Media (cont.) yStorage Medium Where will the information be stored? Storage media - floppy disk, hard disk, tape, CD-ROM etc. yTransmission Medium Over what medium will the information be transmitted? Using information carriers that enable continuous data transmission - networks wire, coaxial cable, fiber optics yInformation Exchange Medium Which information carrier will be used for information exchange between different places? Direct transmission using computer networks Combined use of storage and transmission media (e.g. electronic mail).
Introduction to Multimedia52 Media Concepts zEach medium defines xRepresentation values - determine the information representation of different media Continuous representation values (e.g. electro-magnetic waves) Discrete representation values(e.g. text characters in digital form) xRepresentation space determines the surrounding where the media are presented. Visual representation space (e.g. paper, screen) Acoustic representation space (e.g. stereo)
Introduction to Multimedia53 Media Concepts (cont.) zRepresentation dimensions of a representation space are: ySpatial dimensions: xtwo dimensional (2D graphics) xthree dimensional (holography) yTemporal dimensions: xTime independent (document) - Discrete media Information consists of a sequence of individual elements without a time component. xTime dependent (movie) - Continuous media Information is expressed not only by its individual value but also by its time of occurrence.
Introduction to Multimedia54 Multimedia Systems zQualitative and quantitative evaluation of multimedia systems yCombination of media xcontinuous and discrete. yLevels of media-independence xsome media types (audio/video) may be tightly coupled, others may not. yComputer supported integration xtiming, spatial and semantic synchronization yCommunication capability
Introduction to Multimedia55 Data Streams zDistributed multimedia communication systems xdata of discrete and continuous media are broken into individual units (packets) and transmitted. zData Stream xsequence of individual packets that are transmitted in a time-dependant fashion. xTransmission of information carrying different media leads to data streams with varying features Asynchronous Synchronous Isochronous
Introduction to Multimedia56 Data Stream Characteristics xAsynchronous transmission mode provides for communication with no time restriction Packets reach receiver as quickly as possible, e.g. protocols for transmission xSynchronous transmission mode defines a maximum end-to-end delay for each packet of a data stream. May require intermediate storage E.g. audio connection established over a network. xIsochronous transmission mode defines a maximum and a minimum end-to-end delay for each packet of a data stream. Delay jitter of individual packets is bounded. E.g. transmission of video over a network. Intermediate storage requirements reduced.
Introduction to Multimedia57 Data Stream Characteristics yData Stream characteristics for continuous media can be based on xTime intervals between complete transmission of consecutive packets Strongly periodic data streams - constant time interval Weakly periodic data streams - periodic function with finite period. Aperiodic data streams xData size - amount of consecutive packets Strongly regular data streams - constant amount of data Weakly regular data streams - varies periodically with time Irregular data streams xContinuity Continuous data streams Discrete data streams
Introduction to Multimedia58 Classification based on time intervals Strongly periodic data stream Weakly periodic data stream Aperiodic data stream T T T1T3T2 T1T2 T
Introduction to Multimedia59 Classification based on packet size T D1 T D2 D3 D1 D2 D3 D1 D2 D3 Dn Strongly regular data stream Weakly regular data stream Irregular data stream t t t
Introduction to Multimedia60 Classification based on continuity Continuous data stream Discrete data stream D D1D2D3D4 D D1D2D3D4
Introduction to Multimedia61 Broadband Multimedia Communications Audio/Image/Video Representation
Introduction to Multimedia62 Introduction zBasic Sound Concepts zComputer Representation of Sound zBasic Image Concepts zImage Representation and Formats zVideo Signal Representation zColor Encoding zComputer Video Format
Introduction to Multimedia63 Basic Sound Concepts zAcoustics xstudy of sound - generation, transmission and reception of sound waves. zSound is produced by vibration of matter. xDuring vibration, pressure variations are created in the surrounding air molecules. xPattern of oscillation creates a waveform the wave is made up of pressure differences. xWaveform repeats the same shape at intervals called a period. Periodic sound sources - exhibit more periodicity, more musical - musical instruments, wind etc. Aperiodic sound sources - less periodic - unpitched percussion, sneeze, cough.
Introduction to Multimedia64 Basic Sound Concepts zSound Transmission xSound is transmitted by molecules bumping into each other. xSound is a continuous wave that travels through air. ySound is detected by measuring the pressure level at a point. yReceiving xMicrophone in sound field moves according to the varying pressure exerted on it. xTransducer converts energy into a voltage level (i.e. energy of another form - electrical energy) ySending xSpeaker transforms electrical energy into sound waves.
Introduction to Multimedia65 Frequency of a sound wave period amplitude time Air pressure Frequency is the reciprocal value of the period.
Introduction to Multimedia66 Basic Sound Concepts yWavelength is the distance travelled in one cycle x20Hz is 56 feet, 20KHz is 0.7 in. yFrequency represents the number of periods in a second (measured in hertz, cycles/second). xFrequency is the reciprocal value of the period. xHuman hearing frequency range: 20Hz - 20Khz, voice is about 500Hz to 2Khz. Infrasound from Hz Human range from 20Hz - 20KHz Ultrasound from 20kHz - 1GHz Hypersound from 1GHz - 10THz
Introduction to Multimedia67 Basic Sound Concepts yAmplitude of a sound is the measure of the displacement of the air pressure wave from its mean or quiescent state. ySubjectively heard as loudness. Measured in decibels. 0 db - essentially no sound heard 35 db - quiet home 70 db - noisy street 120db - discomfort
Introduction to Multimedia68 Computer Representation of Audio yA transducer converts pressure to voltage levels. yConvert analog signal into a digital stream by discrete sampling. xDiscretization both in time and amplitude (quantization). yIn a computer, we sample these values at intervals to get a vector of values. yA computer measures the amplitude of the waveform at regular time intervals to produce a series of numbers (samples).
Introduction to Multimedia69 Computer Representation of Audio ySampling Rate: xrate at which a continuous wave is sampled (measured in Hertz) CD standard Hz, Telephone quality Hz. xDirect relationship between sampling rate, sound quality (fidelity) and storage space. xQuestion How often do you need to sample a signal to avoid losing information? xAnswer To decide a sampling rate - must be aware of difference between playback rate and capturing(sampling) rate. It depends on how fast the signal is changing. In reality - twice per cycle (follows from the Nyquist sampling theorem).
Introduction to Multimedia70 Sampling samples Sample Height
Introduction to Multimedia71 Nyquist Sampling Theorem yIf a signal f(t) is sampled at regular intervals of time and at a rate higher than twice the highest significant signal frequency, then the samples contain all the information of the original signal. yExample xActual playback frequency for CD quality audio is Hz xBecause of Nyquist Theorem - we need to sample the signal twice, therefore sampling frequency is Hz.
Introduction to Multimedia72 Data Rate of a Channel yNoiseless Channel Nyquist proved that if any arbitrary signal has been run through a low pass filter of bandwidth H, the filtered signal can be completely reconstructed by making only 2H (exact) samples per second. If the signal consists of V discrete levels, Nyquist’s theorem states: max datarate = 2 *H log_2 V bits/sec noiseless 3kHz channel with quantization level 1 bit cannot transmit binary signal at a rate exceeding 6000 bits per second. yNoisy Channel Thermal noise present is measured by the ratio of the signal power S to the noise power N (signal-to-noise ratio S/N). Max datarate - H log_2 (1+S/N)
Introduction to Multimedia73 Quantization ySample precision - the resolution of a sample value yQuantization depends on the number of bits used measuring the height of the waveform. y16 bit CD quality quantization results in 64K values. yAudio formats are described by sample rate and quantization. Voice quality - 8 bit quantization, 8000 Hz mono(8 Kbytes/sec) 22kHz 8-bit mono (22kBytes/s) and stereo (44Kbytes/sec) CD quality - 16 bit quantization, Hz linear stereo (196 Kbytes/s)
Introduction to Multimedia74 Quantization and Sampling samples Sample Height
Introduction to Multimedia75 Audio Formats yAudio formats are characterized by four parameters xSample rate: Sampling frequency xEncoding: audio data representation -law encoding corresponds to CCITT G standard for voice data in telephone companies in USA, Canada, Japan A-law encoding - used for telephony elsewhere. A-law and -law are sampled at 8000 samples/second with precision of 12bits, compressed to 8-bit samples. Linear Pulse Code Modulation(PCM) - uncompressed audio where samples are proportional to audio signal voltage. xPrecision: number of bits used to store audio sample -law and A-law - 8 bit precision, PCM can be stored at various precisions, 16 bit PCM is common. xChannel: Multiple channels of audio may be interleaved at sample boundaries.
Introduction to Multimedia76 Audio Formats zAvailable on UNIX yau (SUN file format), wav (Microsoft RIFF/waveform format), al (raw a-law), u (raw u-law)… zAvailable on Windows-based systems (RIFF formats) ywav, midi (file format for standard MIDI files), avi zRIFF (Resource Interchange File Format) ytagged file format (similar to TIFF).. Allows multiple applications to read files in RIFF format zRealAudio, MP3 (MPEG Audio Layer 3)
Introduction to Multimedia77 Computer Representation of Voice zBest known technique for voice digitization is pulse-code-modulation (PCM). yConsists of the 2 step process of sampling and quantization. yBased on the sampling theorem. xIf voice data are limited to 4000Hz, then PCM samples 8000 samples per second which is sufficient for input voice signal. yPCM provides analog samples which must be converted to digital representation. xEach of these analog samples must be assigned a binary code. Each sample is approximated by being quantized.
Introduction to Multimedia78 Computer Representation of Music yMIDI (Music Instrument Digital Interface) xstandard that manufacturers of musical instruments use so that instruments can communicate musical information via computers. xThe MIDI interface consists of: Hardware - physical connection b/w instruments, specifies a MIDI port (plugs into computers serial port) and a MIDI cable. Data format - has instrument specification, notion of beginning and end of note, frequency and sound volume. Data grouped into MIDI messages that specify a musical event. An instrument that satisfies both is a MIDI device (e.g. synthesizer) xMIDI software applications include music recording and performance applications, musical notations and printing applications, music education etc.
Introduction to Multimedia79 Computer Representation of Speech xHuman ear is most sensitive in the range 600Hz to 6000 Hz. xSpeech Generation real-time signal generation allows transformation of text into speech without lengthy processing Limited vs. large vocabulary (depends on application) Must be understandable, must sound natural xSpeech Analysis Identification and Verification - recognize speakers using acoustic fingerprint Recognition and Understanding - analyze what has been said How something was said - used in lie detectors. xSpeech transmission - coding, recognition and synthesis methods - achieve minimal data rate for a given quality.
Introduction to Multimedia80 Basic Concepts (Digital Image Representation) yAn image is a spatial representation of an object, a 2D or 3D scene etc. yAbstractly, an image is a continuous function defining a rectangular region of a plane xintensity image - proportional to radiant energy received by a sensor/detector xrange image - line of sight distance from sensor position. yAn image can be thought of as a function with resulting values of the light intensity at each point over a planar region.
Introduction to Multimedia81 Digital Image Representation yFor computer representation, function (e.g. intensity) must be sampled at discrete intervals. xSampling quantizes the intensity values into discrete intervals. Points at which an image is sampled are called picture elements or pixels. Resolution specifies the distance between points - accuracy. xA digital image is represented by a matrix of numeric values each representing a quantized intensity value. I(r,c) - intensity value at position corresponding to row r and column c of the matrix. Intensity value can be represented by bits for black and white images (binary valued images), 8 bits for monochrome imagery to encode color or grayscale levels, 24 bit (color-RGB).
Introduction to Multimedia82 Image Formats yCaptured Image Format xformat obtained from an image frame grabber xImportant parameters Spatial resolution (pixels X pixels) Color encoding (quantization level of a pixel - 8-bit, 24-bit) e.g. “SunVideo” Video digitizer board allows pictures of 320 by 240 pixels with 8-bit grayscale or color resolution. Parallax-X video includes resolution of 640X480 pixels and 24-bit frame buffer.
Introduction to Multimedia83 Image Formats yStored Image Format - format when images are stored yImages are stored as 2D array of values where each value represents the data associated with a pixel in the image. xBitmap - this value is a binary digit xFor a color image - this value may be a collection of 3 values that represent intensities of RGB component at that pixel, 3 numbers that are indices to table of RGB intensities, index to some color data structure etc. yImage file formats include - GIF (Graphical Interchange Format), X11 bitmap, Postscript, JPEG, TIFF
Introduction to Multimedia84 Basic Concepts (Video Representation) yHuman eye views video ximmanent properties of the eye determine essential conditions related to video systems. yVideo signal representation consists of 3 aspects: xVisual Representation objective is to offer the viewer a sense of presence in the scene and of participation in the events portrayed. xTransmission Video signals are transmitted to the receiver through a single television channel xDigitalization analog to digital conversion, sampling of gray(color) level, quantization.
Introduction to Multimedia85 Visual Representation yThe televised image should convey the spatial and temporal content of the scene xVertical detail and viewing distance Aspect ratio: ratio of picture width and height (4/3 = 1.33 is the conventional aspect ratio). Viewing angle = viewing distance/picture height xHorizontal detail and picture width Picture width (conventional TV service ) - 4/3 * picture height xTotal detail content of the image Number of pixels presented separately in the picture height = vertical resolution Number of pixels in the picture width = vertical resolution*aspect ratio product equals total number of picture elements in the image.
Introduction to Multimedia86 Visual Representation xPerception of Depth In natural vision, this is determined by angular separation of images received by the two eyes of the viewer In the flat image of TV, focal length of lenses and changes in depth of focus in a camera influence depth perception. xLuminance and Chrominance Color-vision - achieved through 3 signals, proportional to the relative intensities of RED, GREEN and BLUE. Color encoding during transmission uses one LUMINANCE and two CHROMINANCE signals xTemporal Aspect of Resolution Motion resolution is a rapid succession of slightly different frames. For visual reality, repetition rate must be high enough (a) to guarantee smooth motion and (b) persistance of vision extends over interval between flashes(light cutoff b/w frames).
Introduction to Multimedia87 Visual Representation xContinuity of motion Motion continuity is achieved at a minimal 15 frames per second; is good at 30 frames/sec; some technologies allow 60 frames/sec. NTSC standard provides 30 frames/sec Hz repetition rate. PAL standard provides 25 frames/sec with 25Hz repetition rate. xFlicker effect Flicker effect is a periodic fluctuation of brightness perception. To avoid this effect, we need 50 refresh cycles/sec. Display devices have a display refresh buffer for this. xTemporal aspect of video bandwidth depends on rate of the visual system to scan pixels and on human eye scanning capabilities.
Introduction to Multimedia88 Transmission (NTSC) yVideo bandwidth is computed as follows x700/2 pixels per line X 525 lines per picture X 30 pictures per second xVisible number of lines is 480. yIntermediate delay between frames is x1000ms/30fps = 33.3ms yDisplay time per line is x33.3ms/525 lines = 63.4 microseconds yThe transmitted signal is a composite signal xconsists of 4.2Mhz for the basic signal and 5Mhz for the color, intensity and synchronization information.
Introduction to Multimedia89 Color Encoding yA camera creates three signals xRGB (red, green and blue) yFor transmission of the visual signal, we use three signals 1 luminance (brightness-basic signal) and 2 chrominance (color signals). xIn NTSC, luminance and chrominance are interleaved xGoal at receiver separate luminance from chrominance components avoid interference between them prior to recovery of primary color signals for display.
Introduction to Multimedia90 Color Encoding yRGB signal - for separate signal coding xconsists of 3 separate signals for red, green and blue colors. Other colors are coded as a combination of primary color. (R+G+B = 1) --> neutral white color. yYUV signal xseparate brightness (luminance) component Y and xcolor information (2 chrominance signals U and V) Y = 0.3R G B U = (B-Y) * V = (R-Y) * xResolution of the luminance component is more important than U,V xCoding ratio of Y, U, V is 4:2:2
Introduction to Multimedia91 Color Encoding(cont.) yYIQ signal xsimilar to YUV - used by NTSC format Y = 0.3R G B U = 0.60R G B V = 0.21R -0.52g B yComposite signal xAll information is composed into one signal xTo decode, need modulation methods for eliminating interference b/w luminance and chrominance components.
Introduction to Multimedia92 Digitization yRefers to sampling the gray/color level in the picture at MXN array of points. yOnce points are sampled, they are quantized into pixels sampled value is mapped into an integer quantization level is dependent on number of bits used to represent resulting integer, e.g. 8 bits per pixel or 24 bits per pixel. yNeed to create motion when digitizing video xdigitize pictures in time xobtain sequence of digital images per second to approximate analog motion video.
Introduction to Multimedia93 Computer Video Format yVideo Digitizer xA/D converter yImportant parameters resulting from a digitizer digital image resolution quantization frame rate xE.g. Parallax X Video - camera takes the NTSC signal and the video board digitizes it. Resulting video has 640X480 pixels spatial resolution 24 bits per pixel resolution 20fps (lower image resolution - more fps) xOutput of digital video goes to raster displays with large video RAM memories. Color lookup table used for presentation of color
Introduction to Multimedia94 Digital Transmission Bandwidth yBandwidth requirement for images xraw image transmission b/w = size of image = spatial resolution x pixel resolution xcompressed image - depends on compression scheme xsymbolic image transmission b/w = size of instructions and primitives carrying graphics variables yBandwidth requirement for video xuncompressed video = image size X frame rate xcompressed video - depends on compression scheme xe.g HDTV quality video uncompressed Mbps, compressed using MPEG (34 Mbps with some loss of quality).
Introduction to Multimedia95 Broadband Multimedia Communications Multimedia Compression Techniques
Introduction to Multimedia97 Coding Requirements yStorage Requirements xUncompressed audio: 8Khz, 8-bit quantization implies 64 Kbits to store per second xCD quality audio: 44.1Khz, 16-bit quantization implies storing 705.6Kbits/sec xPAL video format: 640X480 pixels, 24 bit quantization, 25 fps, implies storing 184,320,000 bits/sec = 23,040,000 bytes/sec yBandwidth Requirements xuncompressed audio: 64Kbps xCD quality audio: 705.6Kbps xPAL video format: 184,320,000 bits/sec zCOMPRESSION IS REQUIRED!!!!!!!
Introduction to Multimedia98 Coding Format Examples yJPEG for still images yH.261/H.263 for video conferencing, music and speech (dialog mode applications) yMPEG-1, MPEG-2, MPEG-4 for audio/video playback, VOD (retrieval mode applications) yDVI for still and continuous video applications (two modes of compression) Presentation Level Video (PLV) - high quality compression, but very slow. Suitable for applications distributed on CD-ROMs Real-time Video (RTV) - lower quality compression, but fast. Used in video conferencing applications.
Introduction to Multimedia99 Coding Requirements yDialog mode applications xEnd-to-end Delay (EED) should not exceed ms xFace-to-face application needs EED of 50ms (including compression and decompression). yRetrieval mode applications xFast-forward and rewind data retrieval with simultaneous display (e.g. fast search for information in a multimedia database). xRandom access to single images and audio frames, access time should be less than 0.5sec xDecompression of images, video, audio - should not be linked to other data units - allows random access and editing
Introduction to Multimedia100 Coding Requirements yRequirements for both dialog and retrieval mode applications xSupport for scalable video in different systems. xSupport for various audio and video rates. xSynchronization of audio-video streams (lip synchronization) xEconomy of solutions Compression in software implies cheaper, slower and low quality solution. Compression in hardware implies expensive, faster and high quality solution. xCompatibility e.g. tutoring systems available on CD should run on different platforms.
Introduction to Multimedia101 Classification of Compression Techniques xEntropy Coding lossless encoding used regardless of media’s specific characteristics data taken as a simple digital sequence decompression process regenerates data completely e.g. run-length coding, Huffman coding, Arithmetic coding xSource Coding lossy encoding takes into account the semantics of the data degree of compression depends on data content. E.g. content prediction technique - DPCM, delta modulation xHybrid Coding (used by most multimedia systems) combine entropy with source encoding E.g. JPEG, H.263, DVI (RTV & PLV), MPEG-1, MPEG-2, MPEG-4
Introduction to Multimedia102 Steps in Compression yPicture preparation analog-to-digital conversion generation of appropriate digital representation image division into 8X8 blocks fix the number of bits per pixel yPicture processing (compression algorithm) transformation from time to frequency domain, e.g. DCT motion vector computation for digital video. yQuantization Mapping real numbers to integers (reduction in precision). E.g. U-law encoding - 12bits for real values, 8 bits for integer values yEntropy coding compress a sequential digital stream without loss.
Introduction to Multimedia104 Types of compression zSymmetric Compression Same time needed for decoding and encoding phases Used for dialog mode applications zAsymmetric Compression Compression process is performed once and enough time is available, hence compression can take longer. Decompression is performed frequently and must be done fast. Used for retrieval mode applications
Introduction to Multimedia107 Additional Requirements - JPEG yJPEG implementation is independent of image size and applicable to any image and pixel aspect ratio. yImage content may be of any complexity (with any statistical characteristics). yJPEG should achieve very good compression ratio and good quality image. yFrom the processing complexity of a software solution point of view: JPEG should run on as many available platforms as possible. ySequential decoding (line-by-line) and progressive decoding (refinement of the whole image) should be possible.
Introduction to Multimedia108 Variants of Image Compression zFour different modes xLossy Sequential DCT based mode Baseline process that must be supported by every JPEG implementation. xExpanded Lossy DCT based mode enhancements to baseline process xLossless mode low compression ratio allows perfect reconstruction of original image xHierarchical mode accommodates images of different resolutions
Introduction to Multimedia110 Broadband Multimedia Communications MPEG Compression
Introduction to Multimedia111 Introduction yGeneral Information about MPEG yMPEG/ Video Standard yMPEG/ Audio Standard yMPEG Systems Multiplexing of Video/Audio Data Streams
Introduction to Multimedia112 General Information yMPEG-1 achieves data compression of 1.5Mbps. xThis is the data rate of audio CD’s and DAT’s (Digital Audio Tapes). yMPEG considers explicitly functionalities of other standards,e.g. it uses JPEG. yMPEG defines standard video, audio coding and system data streams with synchronization. yMPEG Core Technology includes many different patents MPEG committee sets technical standards
Introduction to Multimedia113 General Information (cont.) yMPEG stream provides more information than a data stream compressed according to the JPEG standard. xAspect Ratio - 14 aspect ratios can be encoded. 1:1 corresponds to computer graphics, 4:3 corresponds to 702X575 pixels (TV format), 16:9 corresponds to 625/525 (HDTV format). xRefresh Frequency- 8 frequencies are encoded Hz, 24, 25,29.97, 50, 59.94, 60 Hz. yOther Issues with frame rate xEach frame must be built within a maximum of 41.7(33)ms to keep display rate of 24fps(30fps). xNo need or possibility of defining MCUs in MPEG. Implies sequential non-interleaving order. xFor MPEG, there is no advantage to progressive display over sequential display.
Introduction to Multimedia114 MPEG Overview zMPEG exploits temporal (i.e frame-to-frame) redundancy present in all video sequences. zTwo Categories: Intra-frame and inter-frame encoding yDCT based compression for the reduction of spatial redundancy (similar to JPEG) yBlock-based motion compensation for exploiting temporal redundancy xcausal(predictive coding) - current picture is modeled as transformation of picture at some previous time xnon-causal (interpolative coding) - uses past and future reference
Introduction to Multimedia115 MPEG Image Preparation - Motion Representation yPredictive and interpolative coding xGood compression but requires storage and information xOften makes sense for parts of an image and not the whole image. yEach image is divided into areas called macro-blocks (motion compensation units) xEach macro-blocks is partitioned into 16x16 pixels for luminance, 8x8 for each of the chrominance components. xChoice of macro-block size is a tradeoff between gain from motion compensation and cost of motion estimation. xMacro-blocks are useful for compression based on motion estimation.
Introduction to Multimedia116 MPEG Video Processing yMPEG stream includes 4 types of image coding for video processing xI-frames - Intra-coded frames - access points for random access, yields moderate compression xP-frames - Predictive-coded frames - encoded with reference to a previous I or P frame. xB-frames - Bi-directionally predictive coded frames - encoded using previous/next I and P frame, maximum compression xD-frames - DC coded frames yMotivation for types of frames xDemand for efficient coding scheme and fast random access xGoal to achieve high compression rate - temporal redundancies of subsequent pictures (i.e. interframes) must be exploited
Introduction to Multimedia117 MPEG Audio Encoding Steps Psychoacoustic Model Quantization Bit/noise Allocation Filter Bank Multiplexer Entropy Coder Huffman Coding If noise level is too low --> finer quantization is applied If noise level is too high --> rough quantization is applied Transformation from time to frequency domain 32 subbands Compressed data
Introduction to Multimedia118 MPEG/System Data Stream yVideo Stream is interleaved with audio. yVideo Stream consists of 6 layers xSequence layer xGroup of pictures layer Video Param - width, height, aspect ratio, picture rate Bitstream Param - bitrate, bufsize QT - intracoded blocks, intercoded blocks xPicture layer Time code - hours, minutes, seconds xSlice layer Type - I, P, B Buffer Param - decoder’s bufsize Encode Param - indicates info about motion vectors xMacro-block layer Vertical Position - what line does this slice start on? Qscale - how is the quantization table scaled in this slice? xBlock layer