Quyen 1 - Multimedia Foundation
Quyen 1 - Multimedia Foundation
Quyen 1 - Multimedia Foundation
Vic Costello
with Susan A. Youngblood and
Norman E. Youngblood
No part of this publication may be reproduced or transmitted in any form or by any means, electronic or
mechanical, including photocopying, recording, or any information storage and retrieval system, without
permission in writing from the publisher. Details on how to seek permission, further information about the
Publisher’s permissions policies and our arrangements with organizations such as the Copyright Clearance
Center and the Copyright Licensing Agency, can be found at our website: www.elsevier.com/permissions.
This book and the individual contributions contained in it are protected under copyright by the Publisher
(other than as may be noted herein).
Notices
Knowledge and best practice in this field are constantly changing. As new research and experience
broaden our understanding, changes in research methods, professional practices, or medical treatment
may become necessary.
Practitioners and researchers must always rely on their own experience and knowledge in evaluating and
using any information, methods, compounds, or experiments described herein. In using such information or
methods they should be mindful of their own safety and the safety of others, including parties for whom
they have a professional responsibility.
To the fullest extent of the law, neither the Publisher nor the authors, contributors, or editors, assume any
liability for any injury and/or damage to persons or property as a matter of products liability, negligence
or otherwise, or from any use or operation of any methods, products, instructions, or ideas contained in
the material herein.
ISBN: 978-0-240-81394-3
12 13 14 15 16 5 4 3 2 1
For information on all Focal Press
publications
Printed visit ourStates
in the United website
of at
America
ACKNOWLEDGEMENTS..........................................................XViii
ABOUT THE AUTHOR...............................................................XX
ABOUT THE WEBSITE.............................................................Xxi
INTRODUCTION...................................................................Xxii
CHAPTER 9 Text..................................................................245
An Introduction to Typography..............................................245
Legibility and Readability....................................................248
Characteristics that Define Typefaces.....................................249
Stroke, Contrast, and Stress.......................................................249
Weight: Regular, Boldface, and Light...............................................250
Posture: Roman, Oblique, and Italic.................................................250
Proportions and Letterform Parts....................................................250
Categorizing Typefaces......................................................251
Serif Typefaces......................................................................252
From Blackletter to Oldstyle Typefaces...........................................252
Transitional Typefaces...............................................................253
Modern Typefaces.................................................................253
Slab-Serif Typefaces.................................................................253
Sans-Serif Typefaces................................................................254
Decorative Typefaces..................................................................255
Script Typefaces........................................................................255
Contents xiii
Symbol Typefaces and Special Characters.......................................255
Computers and Typefaces.................................................256
Font Styles: True and Faux................................................261
Italics and Boldface for Emphasis.................................................262
Condensed or Expanded Type......................................................262
All Caps and Small Caps.......................................................263
Underline and Strikethrough.......................................................264
Font Size...............................................................................264
Font Color.............................................................................266
Baseline Shift.........................................................................267
Superscript/Subscript..........................................................267
Anti-Aliasing..........................................................................267
Character and Line Spacing...............................................269
Kerning and Tracking................................................................269
Leading..........................................................................269
Alignment, Justification, and Distribution..............................270
Font Transformations......................................................272
Some Final Tips.............................................................274
Limit the Number of Typefaces...................................................274
Text First, Type Second............................................................275
Less Is More..................................................................275
Chapter Summary...........................................................275
CHAPTER 10 Photography........................................................277
Photography............................................................277
Consumer Cameras.............................................................278
Prosumer and Professional Cameras..............................................279
Video Cameras..................................................................280
The Imaging Chain.....................................................280
The Lens.........................................................................281
Prime Lenses..................................................................282
Zoom Lenses..................................................................285
The Iris...........................................................................285
F-stop..........................................................................286
The Shutter.....................................................................287
The Image Sensor..............................................................288
Film Speed....................................................................288
Film Size.............................................................................288
File Size.............................................................................289
Exposure.................................................................292
Measuring Light Intensity...........................................................294
Evaluative Metering Mode........................................................294
xiv Contents
Center-Weighted Metering Mode................................................294
Spot Metering Mode..........................................................294
White Balance..........................................................295
Focus..................................................................297
Autofocus (AF)........................................................................297
AF Target Point....................................................................299
Manual Focus (MF).............................................................299
Depth of Field...............................................................300
Aperture Size.........................................................................300
Focal Length....................................................................302
Subject/Camera Distance..........................................................302
Exposure Modes........................................................302
Automatic Modes...............................................................302
Fully Automatic Mode.............................................................302
Portrait Mode.................................................................302
Landscape Mode..............................................................302
Sports Mode...................................................................303
Night Mode....................................................................303
Semi-Automatic Modes........................................................304
Aperture Priority Mode...........................................................304
Shutter Priority Mode.............................................................305
Program Mode.................................................................306
Manual Mode....................................................................306
Image Stabilization.........................................................306
Downloading Images........................................................308
Digital Image Management.........................................................308
Renaming Images..................................................................309
Image Folders.................................................................309
Chapter Summary...........................................................311
U-Matic ¾-Inch.....................................................................361
xvi Contents
Sony Betacam and Panasonic M............................................362
U-Matic SP.....................................................................362
Betacam-SP and MII................................................................362
Digital Tape Formats................................................................364
D1 and D2...................................................................364
Panasonic D3..................................................................365
Sony Digital Betacam, Betacam SX, HDCAM...................................365
High-Definition Tape Formats.....................................................366
MPEG IMX......................................................................366
Panasonic D5, D5HD..........................................................366
DV Formats......................................................................366
DVCPRO and DVCAM..........................................................366
HDV..........................................................................368
DVCPRO HD....................................................................368
Two Types of Video Compression........................................368
Consumer Formats..........................................................372
Betamax and VHS..............................................................372
MiniDV.........................................................................373
File-Base Recording Formats..............................................374
Open Standards for File-Base Recording.........................................375
MP3..........................................................................376
MPEG-2......................................................................376
MPEG-4 - Advanced Video Coding (AVC).......................................376
Proprietary Standards for File-Base Recording.................................377
Professional Disc and SxS...................................................377
P2............................................................................378
Proprietary Card Formats........................................................378
AVCHD.......................................................................378
Chapter Summary...........................................................381
Full-Screen Graphics................................................................389
Music and Sound Effects.......................................................392
Continuity Editing..........................................................392
Contents xvii
Cut-Ins...........................................................................392
The Project Folder.........................................................394
The Project File......................................................................396
Media Files and Media Clips........................................................396
Capture Files..................................................................398
Render Files...................................................................399
Components of an NLE Interface.........................................401
Project Window......................................................................401
The Timeline..........................................................................401
Video Compositing and Audio Mixing...........................................402
The Preview and Program Monitors...............................................405
Virtual Monitor Components.....................................................405
The Image Frame..................................................................406
Adding Clips to the Timeline..............................................407
A Simple Editing Scenario..........................................................408
Step 1: Set an IN Point and OUT Point in the Preview Monitor............408
Step 2: Set an IN Point in the Timeline........................................409
Step 3: Perform the Edit..........................................................409
Step 4: Repeat Steps 1–3 for SOT #2 and VO #2.........................409
Step 5: Add B-Roll over VO Segments.....................................409
Step 6: Add Titles, Transitions, and Effects...................................410
Transitions......................................................................410
Cut..............................................................................410
Fade............................................................................411
Dissolve.....................................................................411
Wipe............................................................................411
Transition Properties..............................................................412
Chapter Summary...........................................................413
INDEX................................................................................415
Acknowledgements
xviii
INSTRUCTORS
Instructors will find helpful teaching materials on the instructors-only web site.
Please request your login and password by registering here:
http://textbooks.elsevier.com
Introduction
xxii
The idea for this book emerged from a course that I helped create in 2003
and have been teaching ever since. Digital Media Convergence, or DMC as
the stu- dents like to call it, is a rapid-fire survey course that introduces students
to foun- dational concepts and skills used across a wide range of multimedia
technologies and design and production workflows. In DMC, students learn
the basics of visual theory, graphic design, digital photography, sound and
video recording and editing, and Web design—all in a one-semester course. To
accomplish such a feat, this course focuses on breadth rather than depth.
Students gain an expan- sive view of multimedia theory and practice, the
language of audio and visual design, and the essential concepts and principles
they need to understand before moving into advanced courses designed to take
them deeper. One of the biggest challenges in teaching this course has been
finding a textbook that adequately covers all the bases. We’ve looked long and
hard to no avail. And with that, I set out to write one.
Multimedia Foundations: Core Concepts for Digital Design brings together in
one work a myriad of essential principles and concepts common to all forms
of digital artistry. In today’s world of new media and convergent technolo-
gies, the starting point for entry into the digital arts is remarkably the same
whether you aspire to be a journalist or art director, a photographer or film-
maker, a web designer or graphic artist, or a recording engineer or video
producer. The digital revolution established a common denominator for all
human acts of creative expression—namely, bits, those tiny binary morsels
of digital information (0s and 1s) that proliferate invisibly across time and
space bringing countless millions of multimedia messages to the masses
every day.
For better or worse, the computer has become the central tool of multimedia
design and production. Living in a digital age means that all of our devices
and acquisition tools are nothing shy of mini computers designed to capture
the physical elements of the material world (light, sound, motion, etc.) and
convert them into immaterial coded streams of data that a computer or digital
appliance can decipher, process, and manipulate. Your digital camera, smart
phone, iPad, Android tablet, MP3 player, and smart TV serve as extensions
of your digital footprint and toolset. The media you produce and consume
eventually comes home to rest on a physical drive connected to your laptop
or desktop computer, or more frequently today, in the Cloud—on a remote
server housed at an undis- closed location by a multimedia service provider with
a name like Google, Face- book, YouTube, iCloud, or WordPress.
Introduction xxiii
Multimedia Foundations
Section I is intended to give the reader a broad foundational understanding of
the field along with an introduction to some essential tools and processes
used in multimedia design. Chapter 1 sets out to define multimedia. It
explores the characteristics of old and new media from an historical and
theoretical perspec- tive, and it provides an overview of how multimedia
technologies have evolved during key moments of change and innovation.
Chapter 2 looks at the com- puter. No matter what area of multimedia design
you end up pursuing, you will spend lots of time in front of a computer
screen, planning, conceptual- izing, designing, and producing multimedia.
The computer is more than a box filled with high-tech hardware components,
it is an invaluable personal assis- tant, running the tools and programs you
need to create, publish, and consume multimedia. Chapter 3 introduces
readers to the planning and design process. All too often, the planning
process is underestimated, rushed, or eliminated
xxiv Introduction
entirely from the producer’s workflow. Good outcomes are highly dependent
on good planning, vision, and intentionality on the front end of a multimedia
project.
Multimedia Design
Section II includes four chapters on professional practices and theories that
guide and inform the multimedia design process. Chapter 4 draws upon the
field of visual communication, introducing the reader to core elements and
principles of design related to aesthetics, a branch of study dealing with
human perceptions of visual form and presentation. This chapter provides a
conceptual framework to help you objectively assess, discuss, and critique the
presentation of content (or form) within a visual design. Chapter 5 looks at
page design, which has its roots in print media such as newspapers and
magazines. In mul- timedia, the concept of page design extends into the
realm of screen space as well, and digital displays come in a seemingly endless
array of sizes and shapes. This chapter explores concepts related to page
layout and design within the context of two-dimensional digital space.
Chapter 6 looks specifically at inter- face design and the often-related
concept of usability. The multimedia experi- ence is primarily a nonlinear
one in which users rely on the content creator to provide a well-designed
interface to guide them past mountains of irrelevant information to find what
they actually need. You can have the most exciting and compelling content
on your website, but if users can’t find it or get frus- trated trying, they will
click off your site as quickly as they landed. Making the multimedia
experience commonsensical and user-friendly is what this chap- ter is all
about. Chapter 7 looks at web design. While the field of multimedia
encompasses much more than the Web, the ubiquitous Web is what many of
us are most familiar with. For now, it remains the dominant distribution
chan- nel for disseminating multimedia content to the masses. This chapter
explains how the Web works and provides an excellent overview of the
terminology and concepts you need to begin designing, authoring, and
publishing content online via the Web.
Static Media
Section III introduces you to three branches of static media. Static media, as
defined here, refers to media or visual elements that can be presented or viewed
instantaneously, in a fixed moment of time. For example, while you can gaze
endlessly at a photograph, the entire photograph is available to you all at
once. You do not have to wait for it to unfold or reveal itself over time. An
exploration of static media begins with Chapter 8, which takes you into the
field of graphic design and the rudimentary principles used for rendering
raster graphics and bitmap images. Chapter 9 looks at the important field of
typography. Readers will learn concepts related to the placement and design
of text-based informa- tion in a multimedia design. Rounding out this section
is Chapter 10 on digi- tal photography. This is a beautifully illustrated
chapter that covers a wealth of information about camera imaging and
photographic techniques.
Introduction xxv
Time-Based Media
Section IV is concerned with the unique products of time-based media,
namely audio and video, which by definition reveal information to the listener or
viewer linearly over time. Chapter 11 explores the nature of sound and the
production tools and techniques used for capturing and processing digital
audio. Chap- ter 12 explores the historic roots of the audio and video
recording process and the development of analog and digital recording
formats over time. Chapter 13 deals with nonlinear editing (NLE) systems
and their use in professional audio and video postproduction. While
animation is included in most definitions of multimedia as a time-based
element, space did not permit its inclusion in this edition. I intend to cover
the topic of animation in the next edition of this book, as well as online on the
companion website (see page xxi for the website address and instructions).
It’s fun to read and think about what people have said throughout the years
about the Internet and other technologies that impact us daily. Take a moment
and visit the site to see what you can find:
http://www.elon.edu/e-web/predictions/early90s/
Chapter Highlights
Each chapter also includes a list of bullet points that highlights material
exam- ined in the chapter. The bulleted preview section provides a quick
snapshot of the chapter’s content and may be helpful for review as well.
Key Terms
One of my goals for this book is that the reader will gain a greater
understand- ing of the language of multimedia design. The key terms section
features a vocabulary list of important words mentioned in the chapter.
Increasing your
xxvi Introduction
Boxes
Throughout the book, boxes are used to showcase material under three broad
subject headings: Flashback, Great Ideas, and Tech Talk.
FLASHBACK
In order to understand the present, you need to know something about the
past. Flashback segments point to a significant past accomplishment, event, or
individ- ual work.
GREAT IDEAS
The field of multimedia has been affected by countless ideas, inventions, and
innovations that have significantly impacted the way we design and create mul-
timedia. A great idea is not necessarily a good or lasting one, but one that has
left a mark or that has had a transformative effect on the profession and those
who work in it. A great idea may be something as grand as a theory or as simple
as a workflow tip or technique.
TECH TALK
One the biggest challenges I faced in writing this book was deciding how
deep to go into any one particular topic or subject area. Remember, this book
was designed to focus more on breadth than depth. The risk here is that for
some, I will not have gone deep enough and for others I will have gone far
beyond what they were looking for in an introductory textbook. A partial
solution to this dilemma is Tech Talk. Tech Talk delves deeper into advanced
technical concepts and issues associated with multimedia. At times, it also
covers practical tech- niques or applications that go beyond that of a cursory
glance.
SECTION 1
Multimedia Foundations
1. Understanding Multimedia 3
2. The Computer 35
3. Planning and Design 71
Marish/Shutterstock.com
This page intentionally left blank
CHAPTER 1
Understanding Multimedia
We become what we behold. We shape our tools and then our tools Key Terms Algorithm Analog Anim
shape us. Batch Processing Bit
Blog Broadcasting Communication
—Marshall McLuhan, Communication Theorist Hypertext/Hypermedia Hyperconn
Internet Linear
The digital revolution is far more significant than the invention of Mass Audience Mass Media Media
writing or even of printing. Rich Internet Application (RIA) So
—Douglas Engelbart, Inventor of the Computer Mouse
Chapter Highlights
This chapter examines:
Multimedia as an extension of traditional media industries and practices
The five elements of a multimedia experience
Three characteristics of old media
The new media paradigm shift
Five principles of new media in a digital age
WHAT IT IS … IS MULTIMEDIA!
In 1953, legendary comedian Andy Griffith recorded a monologue about a
country preacher’s trip to a college town during a home football game. In this
fictional tale, the preacher has traveled to the “big city” to conduct a tent
meet- ing, but his plans are interrupted when he is unexpectedly caught up by
a fren- zied crowd as they make their way to a football stadium on game day.
What follows is a hilarious first-person account about the culture and sport of
football as witnessed through the eyes of someone who has never seen or
played the game. With a limited vocabulary and frame of reference, he
begins to describe the events around him using the only terms he understands.
He refers to referees as convicts because of their striped uniforms. The
football is called a pumpkin. And the playing surface is compared to a cow
pasture that players enter through a “great big outhouse” on either end of the
field. The skit, titled “What it Was, Was Football,” launched Griffith’s
professional career, leading to a guest appear- ance on the Ed Sullivan show in
1954. The live radio recording remains a cult classic and is one of the biggest
selling comedy recordings of all time.
Multimedia Foundations. DOI: 10.1016/B978-0-240-81394-3.00001-X
© 2012 Elsevier Inc. All rights reserved.
4 SECTION 1 Multimedia Foundations
At oneStructural
Social Media Social Networking time orModularity
another, all of us have been caught by surprise by a new
Text experi- ence or trend that sneaks up on us at lightning speed, challenging old
User-Generated Contentways and habits and leaving us scratching our heads in bewilderment. The
Vlog
country preacher’s first game of football reminds me of the challenge my
Web 1.0/2.0/3.0
World Wide Web mother must have experienced as she learned to send an email message or
open a file attach- ment for the very first time. She was born in the 1930s,
and spent most of her life relying on pen, paper, and the U.S. postal system
for sending and receiving correspondence. To her, this newfangled thing
called email must have seemed like a strange and foreign idea. Perhaps you
can think of a friend, grandparent, or child who has struggled finding the right
words to describe social networking, online shopping, or surfing the Web.
How does someone raised in the 1950s come to understand the World Wide
Web? How does someone raised in the 1970s adapt to Facebook, Twitter,
WordPress, and other social media channels?
For some of you, engaging in a formal study of multimedia will resemble the
country preacher’s first time at a football game. The landscape will appear
strange and foreign to you at first, as you struggle for meaning in a sea of
unfamiliar objects and ideas—even though you’ve probably spent plenty of
time online. In time, a sense of comfort and familiarity will set in as you
catch a glimpse of the big picture and develop a grasp of some fundamental
concepts and rules. To begin, let’s take a peek at something that you are
probably very familiar with that may serve as a common reference point for
understanding multimedia.
Social Media
Social media is a broad term used to describe a growing host of tools and
services that enable computer-mediated interpersonal, group, and mass
communication (see Figure 1.1). Social media can be broken down into many
different catego- ries of services as related to their general purpose and focus.
A few of the most popular channels are included below.
■ Social networking sites such as Facebook, MySpace, LinkedIn, etc., con-
nect people with common interests, backgrounds, and associations. Such
services provide numerous opportunities for synchronous and asynchro-
nous communication through features such as live chatting, email, mes-
sage board posting, event announcements, image galleries, embedded
video, etc.
■ Blogging engines such as Blogger and WordPress provide users with an
online publishing tool for the regular posting of written stories or
narrative commentaries. The term blog is a blended form of the phrase
“web log.” Blogs often focus on a particular subject or offer news and
insight from a spe- cific point of view. They also can serve as a public space
for personal reflec- tions, such as you might find in a diary or travel
journal. Celebrities, media practitioners, and organizations (journalists,
critics, actors, singers, authors, public relations firms, etc.) use blogs for
interacting with fans, consumers, or the general public. Video blogging or
vlogging (pronounced V-logging) is a hybrid form of blogging that uses
Understanding Multimedia CHAPTER 1 5
video in place of a written narrative.
6 SECTION 1 Multimedia Foundations
DEFINING MULTIMEDIA
The word media means “ways of transmission” and encompasses all of the vari-
ous technologies we use to record information and transmit it to others. For
example, videotape is a recording medium (singular) used for storing mov-
ing images and sound onto the physical surface of a magnetic strip.
Television broadcasting and DVD (digital versatile disc) are transmission
media (plural) used to deliver a video recording or live event to an audience.
Likewise, printing is a medium whereby ideas are encoded as letterforms in ink
onto the surface of a page, while books, newspapers, and magazines are the
distribution channels or media through which intellectual content is delivered
to a reader.
Understanding Multimedia CHAPTER 1 7
GREAT IDEAS
The Use of Metaphors
It’s natural for us to draw upon past experiences when confronted with a new tool,
system, or way of doing something. Familiar frames of reference, along with estab-
lished patterns and workflows, can help us to make sense of new technologies and
methods of productivity. This may explain why metaphors are used so often to
describe a new communication object or activity (see Figure 1.2). For example, a
computer’s main interface is called “the desktop” because it represents the
virtual version of a real space where tools and documents reside for conducting
everyday business. Folder icons metaphorically symbolize the place where paper files
are stored in cardboard sleeves. We’re told to think of the hard drive as a file
cabinet. We refer to online con- tent as a “web page” because the book analogy
makes sense to those of us familiar with print media and the structure of content
arranged in a linear format. On Face- book and other social networks we write
messages on “the wall” and we refer to the included members of our social
networks as “friends.” Metaphors are handy devices used to frame complex ideas
in a way that nearly everyone can understand.
FIGURE 1.2
A 1983 promotional brochure from Apple Computer illustrates the power of a good
metaphor. The Apple Lisa computer used these familiar picture icons to represent virtual
versions of common everyday objects in a real office.
Image Courtesy of Computer History Museum.
8 SECTION 1 Multimedia Foundations
GREAT IDEAS
Multimedia
Multimedia is any combination of text, graphics, video, audio, and animation in a distributable format that consumers can in
FIGURE 1.3
Illustrator: Sarah Beth Costello.
Understanding Multimedia CHAPTER 1 9
FIGURE 1.4
USA Today is reaching
out to the next
generation of news
consumers with its
“Mobile Marketplace.”
In addition to
delivering content
through traditional
print and online media
channels, it offers its
readers access to
news through a
growing cadre of
digital devices and
gaming apps.
Source: usatoday.com
ability to rapidly access nearly any form of media content from within the
Cloud or on the Web, through wireless or 3G/4G cellular connections (see
Figure 1.4).
Understanding Multimedia
Today, much of the media content we consume is available in a variety of
formats, intended to serve multiple purposes and audiences. For example,
a book usually starts out as a print-only product. However, if the market
demand is large enough, it may also be published in a spoken-word format
and delivered via compact disc or MP3. With the right equipment, you can
avoid paper altogether by downloading the e-book, a digital version of the
text designed for reading on a computer screen or a handheld device such as
10 SECTION 1 Multimedia Foundations
the Kindle
or iPad. The
website for
a bestseller
may offer
bonus
material or
Understanding Multimedia CHAPTER 1 11
FIGURE 1.5
Visitors to J. K.
Rowling’s official
website (jkrowling.
com) are treated to
an interactive virtual
desktop. Clicking
any object on screen
reveals information
about the author and
her bestselling Harry
Potter books.
Source: jkrowling.com
Tech Talk
FIGURE 1.6
Hypermedia makes
watching a video
online a more
interactive and
nonlinear experience.
Visitors to NASA’s
video gallery
can search the
video library, post
comments, and control
viewing and listening
options using the
media player’s visual
interface.
Source: nasa.gov
Understanding Multimedia CHAPTER 1 13
FLASHBACK
Three Generations of the Web
The evolution of the World Wide Web can be traced through three key stages of
development known as Web 1.0, 2.0, and 3.0. The first generation of the Web,
known as Web 1.0, covers the first decade of its existence from 1991 to 2001.
This era was characterized by one-way communication and point-to-point exchanges
of information. Web pages of this era usually mirrored the linear presentation and
structure of a printed book. Static content, made up mostly of text and images, was
consumed linearly in a traditional manner by reading from left to right and from top to
bottom. User-generated content was unheard of, and there were few opportunities for
interactivity, collaboration, or customization of the user experience. For most people,
access to Web 1.0 was made possible through a low-bandwidth connection using a
dial-up modem. While this was an adequate pipeline for the exchange of text and low-
resolution graphics, it was not sufficient for handling the high-bandwidth transmissions
of large files such as high-resolution images and streaming audio and video files.
Web 2.0 came into its own around 2001 following the dot-com bubble of the late
1990s. This generation of the Web ushered in the era of rich media content, dynamic
web pages, content management systems, content sharing and collaboration sites,
tagging, wikis, blogging, social networks, and more. Web 2.0 was made possible in
large part by the release of program authoring tools used for creating Rich Internet
Applications (RIAs). An RIA is a web-based application such as Adobe Flash, Oracle,
Java, Microsoft Silverlight, or HTML5. RIAs typically require the use of a supported
browser, media player, or browser plug-in (such as Flash Player or QuickTime) to run
programs and view content. RIAs are used for deploying “rich media” or “interactive
multimedia” content to consumers. As the name implies, rich media is designed
to enrich the user’s online experience by increasing opportunities for interactivity,
customization, and personal control of a multimedia experience.
FIGURE 1.7
World Wide Web is the name given to the vast system of interconnected servers
used in the transmission of digital documents formatted using Hypertext Markup
Language (HTML) via the Internet.
(Continued)
14 SECTION 1 Multimedia Foundations
Timothy Berners-Lee, the inventor of World Wide Web, coined the term Semantic Web
to describe Web 3.0, which many see as the next significant iteration of the Web. The Semantic Web is
massive unified database akin to a virtual library. Such a vast library will require more sophisticated co
1 Berners-Lee, T., Hendler, J., & Lassila, O. (2001, May 17). The semantic Web: A new form of Web content that is meaningf
FLASHBACK
My Introduction to the Web
The year was 1993 and I was in a study lounge with fellow graduate students at
the University of Tennessee in Knoxville. We were seated around a small desk
exploring a new software program that had just been installed on a Windows
workstation. The program, called Mosaic (see Figure 1.8), was one of the first web
browsers that could combine colored images with text on the same screen. The more
I discovered about Mosaic’s web-surfing capabilities, the more I began comparing the
experience to that of my parents’ generation, when people gazed upon moving
television images for the very first time. Little did I know then that this nascent
technology would change the world forever and impact me directly in my career as a
video producer and educator.
FIGURE 1.8
Mosaic 1.0 for Microsoft Windows was released in 1993 by the National Center for Super-
computing Applications. NCSA discontinued development and support for the browser in 1997.
Source: NCSA Image Archives.
(Continued)
16 SECTION 1 Multimedia Foundations
Mosaic faded into obscurity nearly as quickly as it came into existence, eclipsed by Netscape Navigato
FIGURE 1.9
These vintage icons
of old media
juxtaposed with a
modern day “play”
button remind us that
the old
media haven’t died,
but rather, they’ve
merely adapted to
new methods of
content delivery and
consumption.
Understanding Multimedia CHAPTER 1 17
FLASHBACK
Paradigm Shift
Thomas Kuhn coined the term paradigm shift in 1962 as a way of describing monumental changes in the meanings of terms
assumptions about natural phenomena. Kuhn was a proponent of aggressive puzzle- solving science, the sort of out-of-the-
A paradigm shift of epic proportions has occurred and things will never be the same.
In the 1950s sociologist Charles Wright examined the mass media and found
that they share three defining characteristics:
1. The mass media are the product of large organizations that operate with great
expense.
2. The mass media are directed towards a relatively large, heterogeneous,
and anonymous audience.
3. The mass media are publicly transmitted and timed to reach the most
audi- ence members simultaneously.
LARGE ORGANIZATIONS
The first characteristic of old media is that they are the product of large organi-
zations that operate with great expense. Hollywood movie studios, metro city
newspapers, recording houses, television networks, broadcast television and
radio stations, and book and magazine publishers, are large entities employing
many people with highly specialized skills and job functions. The concentra-
tion of media ownership has increased significantly over the years, leading to
fewer and fewer companies owning more and more of the world’s mass media
outlets.
In 2010, Fortune Magazine listed The Walt Disney Company as the largest
enter- tainment and media conglomerate in the United States with net
revenues of
$36.1 billion. The top five also included News Corporation, Time Warner, Via-
com, and CBS. These are household names to many of us. Millions of people
are touched daily by content delivered or produced by one of these five com-
panies. The start-up costs for a conventional media operation are high, which
18 SECTION 1 Multimedia Foundations
FIGURE 1.10
Old Media. Seven
industries are often
grouped under the
broad heading of means that most people will never be able to own a television station or
mass media. The print movie studio. Also, professional programming and content is difficult and
media include books, expensive to create. A single episode of a primetime dramatic series can
newspapers, and easily cost more than $10 million to produce.
magazines. Time-
based media (also
known as electronic The Consumer as Producer: The New Media Paradigm Shift
media) include sound Consumers no longer have to rely solely on large organizations to provide
recordings, movies, them with news and entertainment. The shift to new media means that any-
radio, and television.
one can produce and deliver content to a public audience. User-generated
content (UGC), or consumer-generated media (CGM), bypasses the formal
gatekeeping functions and monopolistic control that characterize the old
media factories of cultural production. This paradigm shift is sometimes
referred to as the democratization of media because it empowers the individual
with a multitude of outlets for personal expression. With a little bit of new
media know-how, you can self-publish a book or compact disc, rule a country
in SecondLife, post a video on YouTube, send an iReport to CNN as a citizen
Understanding Multimedia CHAPTER 1 19
FIGURE 1.11
Launched in 2006,
CNN iReport gives
journalist, maintain a blog or Twitter feed, create a website, manage a radio users of CNN.com
station, host a podcast, and so much more. The opportunities for self-expres- a public outlet for
sion and commercial enterprise are virtually unlimited (see Figures 1.11 and posting opinion pieces,
news reports, photos,
1.12). If the content you publish is compelling enough, you could develop a and video. The best
significant following. Almost daily, we hear of people rising to celebrity sta- submissions carry
tus, sometimes almost overnight, by publishing an item of sensational inter- the CNN iReport red
est on the Web. Word of such things travels quickly through the new media stamp of approval,
channels of social networking as well as through mainstream media outlets as meaning they have
been vetted and
they monitor and report on Internet trends and sensations. As Andy Warhol cleared for mainstream
so aptly predicted in 1968, “in the future, everyone will be world-famous for distribution on CNN.
15-minutes.” With new media and the Web, this has never been more likely. Source: ireport.cnn.com
20 SECTION 1 Multimedia Foundations
FIGURE 1.12
Fox News provides the
uReport iPhone app to
citizen journalists for
instant filing of user-
generated photos.
Source: ureport.foxnews
.com
LARGE AUDIENCE
The institutions of old media were optimized to reach a large, anonymous,
and heterogeneous audience. This characteristic identifies the receiver of a
mass media message as a large group of people collectively known as “the
audience.” For this reason, they are sometimes called the “mass audience.”
The model of advertiser-supported media, which has been in use in the United
States since the modern era of commercial printing and broadcasting, makes
content available to consumers for virtually free, or at a partially subsidized
cost. With this model, a media company’s expenses and profits are paid for
indirectly through the sale of commercial advertisements. A “mass” audience
ensures a sufficient return on investment. Television programs will continue or
be discontinued based on their ability to maintain a large following. When
audience numbers fall below a pre- determined threshold or break-even
point, a program is cut to minimize losses.
While market research companies such as Nielsen and Arbitron can provide
aggregate data about the size and composition of a mass audience, the
individ- ual identities of people consuming mass media messages are largely
unknown
Understanding Multimedia CHAPTER 1 21
(see Table 1.1 and 1.2). An exception occurs with subscription services such
as newspapers and magazines. Still, for every known person who subscribes,
there are many anonymous users who acquire print products through point-
of- purchase displays, magazine racks, street vendors, and the like.
The term broadcasting is a metaphor taken from earlier days when farmers
sowed seeds manually by tossing them into the air with a sweeping
movement of the hand. The farmer’s “broadcast” method of planting his
fields ensured that the seed would be evenly dispersed in the air before
hitting the ground. Done cor- rectly, this would produce a healthy crop of
evenly spaced plants. Radio and television broadcasters use this principle to
transmit programming to a mass audience. Based on the frequency and
power allocation awarded to them by the FCC, they have a specific geographic
region in which to operate. The broadcast signal is dispersed evenly over this
area, falling randomly on home receivers that happen to be tuned in at any
given moment. The opportunity to receive a broadcast signal is open to the
public at large. While social and economic factors will prevent some people
from owning a receiver, the wide distribution of the broadcast signal ensures
delivery of the content to a diverse and heterogeneous audience. In order to
appeal to a large and diverse audience, programming has to be broadly
appealing to the average person. While programming can be tar- geted to
specific groups such as men and woman, narrowing the focus too much can
result in an audience size that’s too small to offset expenses or the support of
advertisers who subsidize the cost of over-the-air broadcasting.
Table 1.1U.S. Broadcast TV Ratings for Week of September 19, 2011 Viewers
Rank Program Network Rating (000)
1 Two and a Half Men CBS 16.7 28,743
2 Dancing With the Stars ABC 12.3 19,033
3 NBC Sunday Night Football NBC 12.2 20,357
4 NCIS CBS 12.2 19,958
5 2 Broke Girls CBS 11.5 19,368
6 NCIS: Los Angeles CBS 10.3 16,712
7 Dancing W/Stars Results ABC 9.8 14,787
8 Sunday Night NFL Pre-Kick NBC 9.3 15,975
9 Unforgettable CBS 9.0 14,085
10 Big Bang Theory, the - Sp(S) CBS 8.8 14,939
11 Criminal Minds CBS 8.8 14,137
Media companies and advertisers rely on audience research firms like Nielsen for statistical data about the
consumption patterns of television viewers.
Source: The Nielsen Company. Viewing estimates on this page include live viewing and DVR playback on the
same day, defined as 3:00 a.m. to 3:00 a.m. Ratings are the percentage of TV homes in the United States tuned into
television.
22 SECTION 1 Multimedia Foundations
Time Per
Unique Active Person
Rank Parent Audience (000) Reach % (Hh:mm:ss)
1 Google 373,418 84.8% 2:30:31
2 Microsoft 324,767 73.8% 2:26:48
3 Facebook 280,334 63.7% 6:00:28
4 YAHOO! 235,378 53.5% 1:43:59
5 Wikimedia Foundation 157,527 35.8% 0:13:54
6 EBay 131,274 29.8% 1:28:03
7 Interactive Corp 127,847 29.0% 0:08:46
8 Amazon 122,460 27.8% 0:25:38
9 Apple Computer 115,192 26.2% 1:02:51
10 AOL, Inc. 100,620 22.9% 2:13:44
In response to industry changes, Nielsen has extended its services beyond traditional media analysis into the burgeoning field of new media research
and web analytics.
Source: The Nielsen Company, March 2011.
SIMULTANEOUS DELIVERY
Wright’s second characteristic states that the mass media are publicly transmitted
and timed to reach the most audience members simultaneously. The mass
media industries use expensive distribution systems to ensure that a product is
delivered to consumers in a timely manner. A new motion picture must reach
Understanding Multimedia CHAPTER 1 23
all theaters
24 SECTION 1 Multimedia Foundations
FIGURE 1.13
Pandora Radio provides free access to an unlimited number of virtual radio stations. In addition to choosing from the
list of top stations on the left, users can create their own stations or music channels based on a favorite song, artist, or
genre. With support for mobile apps and social media, Pandora’s reach extends far beyond the Web.
on the targeted release date. The Wall Street Journal arrives in a subscriber’s
mailbox on the publication date whether he resides in Atlanta or Milwaukie.
Television networks go through great pains and lots of research to determine
the best day and time of a week to air their programming. Once the schedule is
set, consumers must align their schedules and expectations to fit the producer’s
timetable.
Old media are also transient in nature. Mass mediated messages are available for
a season, and once they’ve aired or become outdated, they are removed from
circu- lation. Newspapers have a shelf life of only one day for a daily paper or
seven days for a weekly edition. Before the invention of personal electronic
recording devices such as the VCR or TiVo, consumers of electronic media were
Understanding Multimedia CHAPTER 1 25
forced to conform
26 SECTION 1 Multimedia Foundations
to a broadcaster’s fixed program schedule. It was not unusual for families to alter
their lifestyle to accommodate the viewing of a favorite television program.
3
Manovich, L. (2001). The language of new media. Cambridge, MA: The MIT Press.
30 SECTION 1 Multimedia Foundations
NUMERICAL REPRESENTATION
The first principle of new media is called “numerical representation,” and states
that new media objects can be defined numerically as a formal equation or
mathematical function. The computer reduces every act of human
communica- tion to a binary expression made up of zeros and ones.
Old media relied on analog methods of reproduction and distribution. An
analog audio or video recording is represented by a continuous signal, whose
physical qualities vary across time along a linear path. For example, when
listening to a phonograph recording, the stylus is placed in the groove of a vinyl
record. As the record spins, the stylus (usually a diamond or other hard stone)
advances along a spiral path while sensing minute variations in the structure
of the groove. The stylus stays in direct contact with the record at all times,
producing a continuous signal without interruption. The vibrations picked up
by the stylus correspond directly to the fluctuating properties of a live sound
performance across time. The term analog is used to describe this type of
process, because the manmade record- ing is directly comparable to the physical
properties of the original sound source.
The digital sound recording process is different. A digital audio recorder
converts the properties of sound to a numerical value at discrete intervals of
time. Each point of measurement is called a sample. In professional audio
recording, sound is sampled at 48 kHz (or 48,000 times per second). This
means that every 1/48,000th of a second, the computer measures the sound
source’s physical properties and records them as a data file. In a digital
recording, the value of each sample is retained in numerical form, allowing it
to be individually addressed and affected mathematically. While it would be
difficult and impractical to try to edit 48,000 individual samples of a one-
second audio clip, the fact that a digi- tal signal is numerical opens up
opportunities that were impossible to achieve before computerization. New
media, because they are digital, are naturally subject to algorithmic control
and manipulation.
Tech Talk
FIGURE 1.16
Using Adobe Photoshop CS5, various filters were applied to this image of a hurdler to produce a
series of visual effects. The software program performs a mathematical algorithm whenever a
filter, tool, or setting is applied.
Like digital audio recorders, still and motion digital cameras use sampling to
create a digital image. Before computerization, photography involved the
pro- duction of a “continuous” tone image on a sheet of film emulsion. Held
up to a light and magnified, one could notice that the negative imprint of the
image was uninterrupted, a perfect analog of the intensity of light captured by
the pho- tographer through the lens of the camera. Using a digital still
camera, the image plane is broken down into a grid made up of millions of
individual light-sensing picture elements. The computer converts each sample,
or pixel, into a numeric value stored in a data file. Math is an integral part of
the process, and the com- puter uses algorithms to both compress and modify
the pixel values. For exam- ple, painting with the blur tool in Photoshop
causes the pixels in the path of the tool’s brush to lose focus and soften. As
you move the brush across the screen, the image is visibly affected by the
application of the tool. In the background, however, what’s really happening
is that the numerical value of each affected pixel is being changed to reflect
the altered state of its new appearance. As you change the properties of a tool
(e.g., make the brush tip bigger or lighten the stroke intensity, etc.), the
underlying algorithm is adjusted to yield the desired effect (see Figure 1.16).
Convergence
Convergence is a term that’s used often to describe the merging together of
previ- ously discrete technologies into a unified whole. A digital smart phone
is the perfect example of convergence at work in a new media device. The
primary
32 SECTION 1 Multimedia Foundations
FIGURE 1.17
Consumers are relying
more and more on
“multimedia enabled”
smart phones for
24/7 access to Web-
or Cloud-based
services and content
from virtually any
location with a mobile
3G/4G signal.
STRUCTURAL MODULARITY
As we continue our discussion of Manovich’s principles of new media, you will
Understanding Multimedia CHAPTER 1 33
notice that they are somewhat interdependent and often build upon previous
34 SECTION 1 Multimedia Foundations
FIGURE 1.18
In the not-too-distant past, a phone was just a phone. It was used for the sole purpose of
establishing voice communication between two people. Today, the smart phone represents
the convergence of many previously discrete communication technologies. It is a “multimedia
enabled” device that functions more like a computer than the telephone some of us remember
from earlier days.
attributes in the list. Thus, we can say that the second principle of new media,
called modularity, is made possible because of new media’s property of numeri-
cal representation. Modularity means that a new media object retains its indi-
viduality when combined with other objects in a large-scale project. This is only
possible because the computer sees each structural element in a design as a
dis- tinct mathematical object or expression.
The principle of modularity isn’t limited to text alone. To illustrate, let’s take
a look at a common multimedia task that many of you may be familiar with.
A PowerPoint presentation is a good example of structural modularity at
work in the new media age. The digital multimedia slideshow is made up of a
collection of slides positioned and arranged in a logical order of presentation.
The designer can apply a unique transition or timing function to each slide
individually, or he or she may choose to assign a single effect globally to all
of the slides in the presentation. Each slide is populated with its own unique
content and design elements (text, graphics, video, sound, background image,
etc.). Each content element has various properties and attributes associated
with it (position, size, style, etc.). Modularity allows you to build a project
from the ground up, one step at a time, without losing independent control of
each element in the proj- ect. Changes can be made at any time, to any
object, without having to go back to the beginning or to an earlier stage in the
design process.
Understanding Multimedia CHAPTER 1 35
Tech Talk
FIGURE 1.21
The Batch Rename tool in Adobe Bridge CS5 is used to automatically rename the five selected images (top) and copy
them to a new folder (bottom).
38 SECTION 1 Multimedia Foundations
Tech Talk
The software and web-based tools of media creation are growing more
sophisti- cated every year. Prebuilt templates, user-interface components,
widgets, scripts, and style sheets offer users a way to fully or partially
automate complex opera- tions that would normally require extensive
programming expertise to perform. While software tools and web
applications are increasingly more intuitive and intelligent with each new
release, computerized automation has yet to replace the human element.
Producers should strive to embrace the potential efficiencies of workflow
automation without sacrificing the autonomy and aesthetic sensibilities of the
content creator. Some tasks ought never to be delegated to the computer.
VARIABILITY
New media objects are not bound to a single fixed form but can exist, and
often do, in a “potentially infinite” number of versions. When working within
digital workflows, it’s quite common to create and edit projects in one format,
while dis- tributing them in another. For this reason, new documents and
media assets are saved first in the native file format of the application or device
that create them. For example, a Microsoft Word document begins as a .doc or
.docx file. A Photoshop project is saved using the .psd file format. Professional
sound recordings start out as uncompressed WAV or AIFF audio files. And a
digital video project may include source material recorded in a DV, QuickTime,
or MPEG4 compression codec.
Before publishing a Word document online I will usually convert it to a more
widely accessible format such as PDF. PDF is a universal file format that
nearly everyone with a computer and Internet access will be able to download
and view. Having done so, the same document will now exist in two different
formats. I will retain a copy of the MS Word version (.docx) in case I ever need
to make changes to the original document or export it in a different format. The
PDF version will be used only as a distribution format for making the
document available to others.
Likewise, I often use Adobe Photoshop to create graphics for the Web. However,
before publishing a graphic online, it must be saved to a Web-friendly format
Understanding Multimedia CHAPTER 1 39
such as JPEG, GIF, or PNG. To minimize file size and conserve bandwidth
during transmission, online music files and audio podcasts are distributed in
the popu- lar MP3 file format. As a final example, in order for a video project to
be burned
40 SECTION 1 Multimedia Foundations
FIGURE 1.22
This image of a red
boat was saved first
in Adobe Photoshop’s
native PSD file format.
However, the program
offers lots of options
for saving a
document file in a
different image
format.
Understanding Multimedia CHAPTER 1 41
CULTURAL TRANSCODING
Manovich’s fifth principle of new media delves deeply into the theoretical impli-
cations of using computers to represent the various products of human
culture. He uses the term cultural transcoding to describe the bi-directional
influence of computers and human culture acting reciprocally upon one
another. With old media, the methods of production and distribution were
based largely on organic processes that conformed to the natural properties of a
specific medium and method of reproduction. Sound and light were captured
and stored in a way that preserved the continuous structure and form of the
content in a natu- ral way. Manovich believes that new media is partitioned
into two competing parts, what he calls the “cultural layer” and the
“computer layer,” and that each layer is constantly at work influencing the
other. The merging of the two layers is producing a new “computer
culture”—“a blend of human and computer mean- ings, of traditional ways in
which human culture modeled the world and the computer’s own means of
representing it.” In the words of Marshall McLuhan, “We become what we
behold. We shape our tools and then our tools shape us.”4
In the mid-1990s, Nicholas Negroponte, then director of MIT’s Media Lab, argued
CHAPTER SUMMARY
that the world’s economy was moving rapidly from the production of material objects
com- prised of atoms to immaterial ones made of bits. He knew that any object of
human cre- ativity that could be digitized could also, then, be delivered to consumers
via the world’s growing networks of computer and telecommunication systems. Digital
platforms would become the dominant means for designing, producing, and distributing
ideas in the next information age. Less than twenty years later, we know that this is
true. As of 2011, Amazon now sells more e-books per year than physical books;
YouTube is the largest distributer of online video content in the world; and if Facebook
was a country, we’re told that based on users it would be the third-largest nation in
the world. This move from physical to digital production and consumption has
rapidly reshaped how people com- municate and how users consume the media
products that professionals create. We are no longer satisfied consuming media on
an à la carte basis, especially at the whim of producers using proprietary methods
and fixed delivery systems. With new media, the consumer is potentially just as
influential as the producer. In fact, we can no longer speak in such archaic ways about
consumers and producers. In the multimedia age, everyone can be both and often
is.
In this first chapter, we have defined multimedia as any communication event that
involves the combination of text, graphics, video, audio, and animation though a digital
channel or device. Understanding multimedia requires an appreciation for the legacy
media and systems that have influenced human communication for more than a
hundred years. Traditional television, books, newspapers, etc. haven’t left us, per se; but
rather, they are changing and evolving under a radically new paradigm and production
rubric. Likewise, we need to understand what has been added to our knowledge as a
result of new media. The digital revolution brings with it a whole new set of concepts
and workflows that must be learned and understood. As with many things, understanding
multimedia involves looking forward to the future through the lens of past
experience and practice.
4
McLuhan, M. (1964). Understanding media: The extensions of man. New York: McGraw-Hill.
This page intentionally left blank
CHAPTER 2
The Computer
35
Chapter Highlights
This chapter examines:
Personal computing and the digital revolution
The role of computer hardware and software as a tool for the multimedia producer
The factors affecting the speed and efficiency of computer processing
The Human Interface Device—using the mouse and keyboard to interact with a computer
The Graphical User Interface—using visual prompts and symbols to interact with a computer
Storage solutions for saving, retrieving, and managing digital files and project assets
MS-DOS
Network- Attached Storage Operating System OS X
Table 2.1Decimal to Binary
PC QWERTY RAM
0 0
1 1
2 10
3 11
Removable 4 100
Storage 5 101
Serial Bus
Software 6 110
Terabyte 7 111
USB
Windows 8 1000
9 1001
10 1010
converts the decimal numbers 0–10 into binary, and you’ll notice that the
number 2 in binary is 10.
What’s that you say? Why is this important in a book on multimedia production?
Good question! It’s important because the binary system is a critical component
of the computer. When researchers were looking at ways to use electricity to do
mathematical calculations, they realized that they could harness the fact that
an electrical circuit was either on or off, if they used binary rather than
decimal numbers—an idea very similar to that used by Samuel Morse with the
telegraph. In a computer, each instance of a zero or one is called a bit. Binary
numbers (bits) are still at the heart of how the computer stores and processes
information, even if contemporary computers shield us from them. Bits are a
critical component of what Nicholas Negroponte talked about in the early
1990s when he argued the economy was making a transition from atoms to
bits, the move from physical items such as paperback books to digital items,
such as Amazon’s Kindle e-book reader. Part of the process of becoming a
multimedia professional is learning the basics of what is happening inside the
box of electronics that sits on or near your desk, and what each of the parts do.
Understanding this will help you in select- ing the appropriate tools for a job,
and being able to talk coherently about the technology increases your
credibility and your chances for employment.
In the early 1960s, the American Standards Association began work on ASCII
(pronounced as-KE), the American Standard Code for Information Inter-
change, to provide an industry standard and make it easier to move data
between computer systems. ASCII translates its 128-character set, including
most of the symbols found on a typewriter keyboard, into a binary format
(see Table 2.2). While ASCII and regional variants have dominated the com-
puting industry, it is gradually being replaced by UCS Transformation—8 bit
(UTF-8), particularly on the Web, though the UTF-8’s first 128 characters are
the same as those in ASCII.
Using the ASCII character set, the phrase “smiley face” can be translated into
binary code as 11 bytes (or 88 bits) of digital data (see Table 2.3).
As you can see, it takes a lot of zeros and ones to represent a very simple
two- word phrase. The more complex the information, the more bits and bytes
will be needed to encode it into a form that a digital device can recognize.
Table 2.3 The Phrase “Smiley Face” Represented with Binary Code Consumes 88 Bits
s m i l e y
01110011 01101101 01101001 01101100 01100101 01111001
space f a c e
GREAT IDEAS
Communicating with Binary Numbers
He said to his friend,—“If the British
march By land or sea from the town to-
night,
Hang a lantern aloft in the belfry-arch
Of the North Church tower as a signal
light,— One if by land, and two if by sea;
And I on the opposite shore will be,
Ready to ride and spread the alarm
Through every Middlesex village and
farm, For the country folk to be up and
to arm.”
Henry Wadsworth Longfellow, “Paul Revere’s Ride” (1861)
Longfellow’s classic line, “one if by land, and two if by sea,” illustrates the
usefulness of a binary code for communicating a message. On the eve of the
American Revolutionary War, Paul Revere and his associates developed a surefire
plan for
The Computer CHAPTER 2 39
warning the patriots of the British army’s route of attack. If the British approached by land, a single lantern was to be hun
of the telegraph the next century turned binary signaling into an international communication tool.
The way that a computer processes digital information today is similar to how a human telegraph operator would have en
the 1830s, Morse code is a binary encoding scheme that uses dots and dashes to represent letters, numbers, and punctuat
Each character in a Morse code transmission is a carefully crafted series of short and long electrical bursts. For example, o
FIGURE 2.2
A telegraph key such as this one was used to encode messages in Morse code, an early system
of binary transmission using dots and dashes.
40 SECTION 1 Multimedia Foundations
FIGURE 2.3
A large space was
required to house the
massive components
of the ENIAC at the
U.S. Army Research
Lab in Adelphi,
Maryland.
Source: K. Kempf, U.S.
Army Research Lab.
The Computer CHAPTER 2 41
(see Figure 2.4). While MITS hoped to sell a few hundred systems the first year,
it was deluged with thousands of orders following an advertisement in
Popular Electronics. Microsoft (originally MicroSoft), founded by Paul Allen
and Bill Gates, got its start by developing software for the new Altair. The next
year, two friends in California, Steve Wozniak and Steve Jobs, formed a
small computer company called Apple to market a kit-based computer, the
Apple I. The next year, they introduced the Apple II, a fully functional
computer. The age of per- sonal computing had arrived.
In 1981, pushed in part by the success of Apple and others, IBM introduced
the IBM 5150—the IBM PC. It sold well, buoyed in part by IBM’s reputation
for mainframe computing and by IBM’s decision to use “open architecture”
hardware, which meant that technical specifications were made public, encour-
aging other companies to build add-ons. This also allowed other companies to
find ways to build “clones,” systems that mimicked the IBM PC. IBM decided to
have another company, Microsoft, develop the operating system (OS) and let
Microsoft retain ownership. This decision meant that Microsoft could license
its OS, MS-DOS (MicroSoft-Disk Operating System) to other companies.
Microsoft’s retention of the right to market MS-
DOS helped propel the company to international
fame, and the abbreviation PC evolved into a
generic nickname for many computers based on
IBM’s original design. The social and economic
impact of the microcomputer was felt around the
world, securing it as one of the most influential
technolog- ical breakthroughs of the 20th century.
Less than 10 years after the introduction of the
microcomputer, the newsmagazine Time selected FIGURE 2.4
the computer as its “Man of the Year” for 1982. The Altair 8800
Microcomputer was
released in 1975.
Source: Ed Uthman,
Creative Commons
License BY-SA-2.0, via
Wikimedia Commons.
FIGURE 2.5
Computer scientist Alan Kay created this mockup prototype of the Dynabook tablet computer
(left) in 1968. While tablet computers like the iPad are a relatively recent phenomenon, the idea
for them was hatched over 40 years ago. On the right, students explore the visual interface of
the Alto PC, developed at Xerox PARC in 1973. The Alto was the first desktop computer to utilize
a graphical user interface and mouse.
Source: PARC Media Library.
42 SECTION 1 Multimedia Foundations
FIGURE 2.6
This collection of
magazine covers Two years later, Apple introduced the Macintosh, simplifying the way users
from 1978 to 2010 interact with the computer and helping Apple to solidify a profitable niche in
chronicles consumers’ the burgeoning desktop publishing industry, a professional sector that
infatuation with
computer technology Apple’s advancements in computers and laser printers helped fuel. Apple’s
over three decades of commitment to desktop publishing, and its subsequent foray into desktop
product innovation. video editing, fos- tered a loyal base of support among multimedia
professionals. Except for a brief stint in the 1990s, Apple has refused to license
its operating system to other manufacturers of desktop and laptop computers.
this process “booting up.” The act of restarting a computer (or “rebooting” it),
repeats the boot sequence—refreshing the operating system and restoring the
system to its default configuration. Other software, or applications, are designed
to perform specific tasks. Be aware that some of the software you use as a mul-
timedia professional may not be on your computer, it may be on the Web. This
trend includes distribution tools such as YouTube but also, increasingly, produc-
tivity tools ranging from word processors and spread sheets such as Google Docs
to online photo-editing tools such as Photoshop Express.
It may be helpful to think about hardware and software as you would think
about a book. The physical substance of a book—the binding, cover, pages, and
ink—are like hardware, while the intellectual substance of a book, encoded
in
Tech Talk
FIGURE 2.7
A computer is a carefully engineered
collection of individual components.
The Computer CHAPTER 2 45
the form of words, is like the software, providing meaning and significance
to the reader. In short, a computer without software is about as meaningful as
a book filled with empty pages.
Windows Macintosh
Intel® Pentium® 4 or AMD Athlon®
64 processor Multicore Intel Processor
requirements. If possible, you’d want substantially more RAM than what Adobe
lists as the minimum, particularly if you are working with large images and if
you plan to have any other software applications running at the same time as
Photoshop.
PROCESSOR SPEED
Located prominently on the computer motherboard is a large silicon chip called
the central processing unit (CPU), or more simply, the processor (see Figure
2.8). It serves as the brain of the computer and plays an important role in
determin- ing the overall speed and efficiency of the system. Intel is the leading
manufac- turer of microprocessors.
Chip manufacturers can improve the performance capabilities of
microproces- sors in several ways. First, they can increase the microprocessor’s
transistor den- sity, the physical number of transistors on the chip’s surface.
Second, they can increase the rate at which a processor performs basic
operations—the processor’s “clock speed,” the rate at which a processor can
execute instructions. Clock speed is measured in millions of cycles per second,
expressed as megahertz (MHz); or in billions of cycles per second, expressed
as gigahertz (GHz). It’s important to understand that the processor with the
highest clock speed is not always the fastest processor. Some processors have
more efficient instruction sets than oth- ers, and manufacturers are building
multicore CPUs, CPUs that have multiple independent processors (or cores)
on a single chip. Dual-core (2), quad-core (4), and 8-core processors are now
quite common. It appears that the old adage “two heads are better than one”
applies to the brainpower of a computer as well. In order to take advantage of
multicore processing, operating system and applica- tion software must be
written to support it.
FIGURE 2.8
Don’t be fooled by its small size. The CPU is
a silicon chip or wafer containing millions of
tiny switches called transistors. It functions
as the brain of a computer, performing all of
the mathematical operations required of the
operating system and program software.
The Computer CHAPTER 2 47
FLASHBACK
Moore’s Law
In 1965, Intel cofounder, Gordon Moore, predicted that the number of transistors
per square inch on integrated circuits (the predecessor of the microprocessor)
would double about every two years for the foreseeable future. The prediction, now
referred to as “Moore’s law,” has proved markedly accurate. Intel introduced its first
microprocessor, the 4004, in 1971. It used 2300 transistors. The Intel 8088 (1979)
used in the original IBM PC, had 29,000. The Pentium III processor was introduced in
1999 and contained 9.5 million transistors. In 2010, Intel announced release of the
Itanium 2 processor, a family of multicore chips with more than 2 billion transistors.
FIGURE 2.9
Gordon Moore (pictured left) accurately predicted the exponential growth in the number of
transistors on a CPU.
Source: Intel
The chart (pictured right) illustrates Moore’s law across the span of four decades.
Source: Wgsimon. Creative Commons License BY-SA-3.0, via Wikimedia Commons
SYSTEM MEMORY
Memory is a critical part of the computer. It typically falls into one of two
cat- egories. The first type is memory used for data storage, such as hard
drives, which are nonvolatile, meaning that they retain information even
when the computer is off. We’ll be talking about that type of memory later in
the chapter. The second type, and the topic at hand, is system memory, often
called RAM (Random Access Memory). RAM is used to temporarily store
operating system software and program files while a computer is running.
This type of memory is volatile memory, which means that data is retained in
RAM only as long as an electrical current is provided to the memory chips.
System memory is erased (or refreshed) each time the computer is turned off
or restarted.
When a program like Photoshop CS5 is launched, the files necessary to run
the program are loaded into RAM and will remain there until the program
48 SECTION 1 Multimedia Foundations
shuts
The Computer CHAPTER 2 49
FIGURE 2.10
A RAM chip is carefully inserted into a designated slot on the computer’s motherboard.
down (the user quits or exits the application). Computers do this because
read- ing data from RAM is much faster than reading it directly off of the
hard drive. The speed of RAM is rated in nanoseconds (billionths of a
second), while hard drive speed is rated in milliseconds (thousandths of a
second). As we’ll see in a minute, this is one of the reasons that it is critical
to make sure your computer has plenty of RAM.
Manufacturers preconfigure computer systems with a minimum amount of sys-
tem memory. This is like selling a car with only standard features. Upgrades are
available, but they come at an additional cost. Editing and design programs tend
to require more system memory than programs like Internet browsers and word
processors (see Table 2.5).
Software memory recommendations do not factor in the likelihood that users
will work with more than one application at a time. The more applications
you have open on your computer at one time, the more RAM you will need
to run the software. At some point, you will simply run out of RAM and not be
allowed to open any additional programs until you begin shutting things
down. As a general rule, it’s better to err on the side of having more RAM
than you think you’ll ever need than having to live with sluggish
performance and perpetual system crashes brought about by not having
enough system memory. Adding more RAM to a computer system is one of the
best things you can do to increase the speed and performance of software
applications. One of the reasons adding RAM helps is that the less RAM you
have, the more often the computer has to write information to the hard drive
and writing information to and from the hard drive is much, much slower
than writing it to RAM.
50 SECTION 1 Multimedia Foundations
As you read system requirements for software packages, you’ll also notice that
some of them talk about specifications for the video card—the piece of hardware
that gets information from the computer to the computer screen (see Figure
2.11); we’ll talk more about the screen itself in Chapter 8, “Graphics.” As
graphics became a more and more important part of the computing experience,
particularly with graphics- intensive video games, computer designers began
including graphics processing
Windows 7 OS 1–2 GB
Mac OS X (Lion) 2 GB
Mozilla Firefox 128 MB
Microsoft Office for Mac 2011 1 GB
Adobe Flash Professional CS5 2 GB
Pro Tools 9 4 GB
Final Cut Pro X 4 GB
Avid Media Composer 4 4 GB
FIGURE 2.11
A video card (also known as a video adapter or graphics card ) like this one is attached
directly to the computer’s motherboard. The video signal is fed to a computer monitor via a
cable connected to an output connector on the outside of the box.
The Computer CHAPTER 2 51
units (GPUs) on video cards to take the graphic processing load off of the CPU.
Bet- ter video cards will have their own RAM, separate from what’s on the
motherboard. Programs that put a strain on video cards, such as video editing and
3D modeling, often have minimum requirements for the GPU and for the amount
of RAM avail- able on the video card. Some will even require a specific brand of
video card.
FIGURE 2.12
The evolution of the
computer mouse.
(A)The original mouse
prototype invented by
Douglas Englebart in
1963.
Source: SRI International.
(B)The first
commercial three-
button mouse
developed by Xerox
PARC.
Source: PARC Library.
(C)A modern wireless
mouse.
FIGURE 2.13
Apple produced this marketing piece in 1983 to highlight Lisa’s user-friendly graphical user
interface (GUI). Image courtesy of Computer History Museum.
54 SECTION 1 Multimedia Foundations
Windows caught up with the Macintosh. Today, the GUI is the norm for
most operating systems, even if CLI-level access is still available.
The Keyboard
Despite the importance of the mouse and other input devices such as touch-
based screens and tablets, the keyboard is still one of the main input devices
used for entering data into the computer. While there is no universal standard
for the layout of a computer keyboard, most English language keyboards use
the QWERTY (pronounced KWER-tee) arrangement for the layout of the alpha-
numeric keys (see Figure 2.14). So named because the top row of alphabetic
characters begins with the letters q, w, e, r, t, y, the keyboard was designed in
1878 to space out the most commonly used keys to prevent the typing
mecha- nism from jamming. While jamming is not an issue with electronic
keyboards, widespread familiarity with QWERTY keyboards has limited the
adoption of other systems. Although similar in appearance to typewriter
keyboards, com- puter keyboards have more keys with specialized functions,
and the user can often redefine how specialty keys work. For example, in
text-oriented applica- tions, the space bar is used to insert a space between
characters. However, in most video editing programs, the space bar also
functions as a play/pause con- trol for starting and stopping video segments.
KEYBOARD BASICS
There are three main categories of keys on a computer keyboard: character keys,
modifier keys, and function keys.
FIGURE 2.14
The QWERTY layout
dates back to the 1800s
and is still used today
on English language
keyboards.
The Computer CHAPTER 2 55
FIGURE 2.15
The Character keys
consume most of
the space on a
standard computer
Character Keys
keyboard. Character keys are used to insert letters, numbers, and punctuation symbols into
a document or text-box window (see Figure 2.15). They also include several
related keys used for document formatting and data entry.
■ Space Bar. The space bar is most commonly used to insert a space
between characters and words. In audio and video applications, it is often
used to stop and start playback.
■ Tab Key. In word-processing applications, the Tab key is used to move
the insertion point for text to the next predefined tab stop located along the
hori- zontal ruler at the top of the document. Tab stops generally occur at
regular intervals such as every half-inch. In non-text-oriented
applications, the Tab key can be used to advance the location of the cursor
to the next item in a sequence. For example, when completing an online
form in a web browser, selecting the Tab key will advance you to next entry
field. The Tab key is also used for advancing from cell to cell in a table or
spreadsheet.
■ Return/Enter Key. When working with text, the Return key is used to advance
the insertion point to the beginning of the next line. When performing
data entry, the Return key completes the input of data in one cell and
moves the insertion point to the next entry field. The Enter key is located
on the numeric keypad and in most cases functions the same as the Return
key.
■ Backspace and Delete Keys. Depressing either the Backspace or Delete keys
will remove selected text or objects from a project. In this case, both buttons
function pretty much the same. However, when typing text, the Backspace
key deletes characters located before the insertion point, while the Delete key
removes characters located after the insertion point. The Backspace key may
be labeled backspace, delete, del, or simply have an icon of a backward-
facing arrow. The Delete key may also be labeled delete or del, or have an
icon of a forward-facing arrow.
Modifier Keys
Modifier keys are used to alter the actions of other keys or mouse clicks
56 SECTION 1 Multimedia Foundations
(see
Figure
2.16).
The Computer CHAPTER 2 57
keys let
you move
to the top
or bottom
of a
document.
■ Shift Key. The Shift key is used to change the display of a character to its In text
upper- case form or to select the upper character of a dual-purpose key such editing
as using SHIFT+5 to create the percent symbol (%). The Caps-Lock key
forces all letters to appear as capitals, though it doesn’t lock the other keys
the same way.
■ Control/Command Keys. The Control key on a PC and the Command key on
a Mac usually function the same way. Older Mac keyboards may have an
Apple on the Command key. Command is used most often in combina-
tion with other keys to provide shortcuts for menu item commands. For
example, on a PC, CTRL+S (CMD+S on a Mac) is usually the keyboard
shortcut for saving the current file.
■ Alt/Option Keys. In Windows, the Alt key is used in combination with let-
ters to open the drop-down menus on the menu bar. For example, select-
ing ALT+F opens the File menu; ALT+E opens the Edit menu; etc. On an
Apple keyboard, the Alt key is called the Option key, but it behaves rather
differently than the Alt key on a PC. When combined with other keys, the
Option key can be used to generate special text symbols. For example, select-
ing OPTION+G on a Mac produces the international copyright symbol (©).
The Option key can also be used to modify the effect of a mouse click or
drag.
■ Escape Key. Originally designed to cancel or abort a running command or
procedure, the Escape key is used most often in combination with other keys
and less often as a standalone key function.
Function Keys
Function keys are dedicated to performing specific tasks (see Figure 2.17).
■ F-keys. Programmable function keys labeled F1, F2, etc. are located across
the top row of the keyboard. These keys perform different functions
depending on what type of computer you are working on and what
application you are currently working in. Many applications allow the user
to assign frequently- used-tasks to specific function numbers to speed up the
process of executing repetitive actions.
■ Home/End. In some programs, such as Adobe Acrobat, the Home and End
58 SECTION 1 Multimedia Foundations
FIGURE 2.16
Modifier keys are used in combination with other keys to alter their function.
The Computer CHAPTER 2 59
FIGURE 2.17
Function keys may
perform specific tasks
as assigned by the
user or the system
operating software
or computer program
being used.
programs such as Microsoft Word, home and end will move the insertion
point to the beginning or end of a sentence. In video editing programs
such as Final Cut Pro, the Home key will move the play head in the
editing time- line to the beginning of the program. The End key will move
it to the end of the timeline. Did we mention some keys will work
differently depending on what program you are using?
■ Page Up/Page Down. Depressing Page Up or Page Down scrolls the
viewable area of a document or a window up or down one page or screen at
a time.
■ Arrow Keys. In text editing, arrow keys can be used to reposition the
inser- tion point to the left or right one space at a time, or up and down
one line at a time. When used with the Shift key, arrow keys can be used
to expand or collapse the selection of words, sentences, and paragraphs. In
graphic design programs such as Photoshop, the arrow keys can be used to
nudge the posi- tion of an object one pixel at a time in either direction. In
tables and menus, the arrow keys are used to move horizontally and
vertically from cell to cell or item to item.
Keyboard Shortcuts
Keyboard shortcuts can help save time and improve workflow productivity.
System-specific shortcuts (see Table 2.6) are built into the operating system
software and will normally work with any platform-compatible software
application, though this is not always the case. Application-specific shortcuts
(see Table 2.7) are designed to work only within a specific software
application and will vary from program to program. A convenient way to learn
about shortcuts is to explore the program’s drop-down menus. If a shortcut for
a specific action is available, it will usually be displayed to the right of the
command line description.
Keystroke Combinations
Keystroke combinations are typically written in an annotated style where the key
name or abbreviation is connected to another key name by either a hyphen (-)
or plus sign (+). For example, the infamous Windows shortcut
CTRL+ALT+DEL instructs the operator to depress and hold the Control key,
followed by the Alt key, followed by the Delete key, until all three keys are
simultaneously depressed. In Windows XP, the CTRL+ALT+DEL keystroke is
60 SECTION 1 Multimedia Foundations
used to end a task or exit from an
The Computer CHAPTER 2 61
CTRL+O
CMD+O Open File
CTRL+N
CMD+N New File or New
Finder Window
CTRL+P
CMD+P Print Save
CTRL+S
CMD+S Close File
CTRL+W
CMD+W Exit/Quit Current
CTRL+X
CMD-Q Application
Switch between open applications
ALT+Tab
CMD+Tab Select All
CTRL+A Find
CMD+A Copy Cut
CTRL+F
CMD+F Paste
CTRL+C
CMD+C Undo
CTRL+X
CMD+X
CTRL+V
CMD+V
CTRL+Z
CMD+Z
Right-Click Right-Click or Opens Contextual
CTRL+Click (with a Menu
1-button mouse)
CTRL+ +
Zoom In Zoom
CTRL+ -
Out
V
Activates the Move tool Free
CTRL+T
Transform
space on a digital storage device. The more work you generate with a com-
puter, the more files you will generate, and the more options you will need
for managing, storing, retrieving, and moving files. Computer storage devices
are rated in terms of how much binary data they can hold. In 1998, the
International Electrotechnical Commission (IEC) approved the following
units of measurement for system memory chips and computer storage devices:
kilobyte (KB) = 210 or 1,024 bytes of data
megabyte (MB) = 220 1,048,576 bytes of data
gigabyte (GB) = 230 or 1,073,741,824 bytes of data
terabyte (TB) = 240 or 1,099,511,627,776 bytes of data
petabyte (PB) = 250 or 1,125,899,906,842,624 bytes of data
exabyte (EB) = 260 or 1,152,921,504,606,846,976 bytes of
data
While these base-2 equivalents are technically correct, the practice of
rounding off has become commonplace for consumers and manufacturers
alike. With numbers this large, many people find it easier to think of a
Tech kilobyte
Talk
FIGURE 2.18
The file system browser of a computer functions virtually the same, regardless of whether you are using a PC or a Mac.
On a PC, Windows Explorer (left) is used to access and manage applications, folders, and files. On a Mac, the Finder
application performs many of the same functions.
64 SECTION 1 Multimedia Foundations
FIGURE 2.19
The file system browser gives users many options for conducting advanced searches. For example, the Finder in Mac OS X allows you to
FIGURE 2.20
A comparison of data storage requirements for different types of multimedia content. Audio and video files consume
much more space on a digital storage drive than simple text and graphics.
Tech Talk
File Management 101 One of the worst ways to pack a containing related
storage building is to randomly toss items into the unit
one at a time until nothing else can fit. Doing so will
create a mess and make it extremely difficult for the
owner or any- one else to locate and retrieve specific
items at a later time. You’ll also waste a lot of precious
space in the process. A savvy packer will approach this
task more thoughtfully. For example, one could begin by
packing items into boxes along with other items of a
related nature. Each box could then be labeled
according to its contents (kitchen items, toys, crafts,
tools, books, Christmas decorations, etc.).
This kind of organizational system has been used for
years to manage physical documents. For example, a
doctor keeps tabs on your medical history by updating
your file each time you come in for a checkup. Your
medi- cal file is placed in a folder labeled with your
name or patient ID number. The folder is stored
alphabetically or alphanumerically in a specific drawer
or section of the file cabinet. Like the boxes in our
storage facility, the folder serves as a physical
container keeping all of your information together in
one place and separate from that of other patients.
Proper labeling and sorting helps the office staff to
easily locate your file among the thousands of identical
folders in the cabinet.
Because this system was so widely understood, it was
adopted metaphorically for managing data on computer
storage drives. Thus, any digital storage device, such as a
hard drive, flash drive, memory card, etc., can be
thought of as a virtual file cabinet designed for holding
electronic documents. To avoid clutter, files are
placed strategi- cally into named folders or subfolders
66 SECTION 1 Multimedia Foundations
FIGURE 2.21
Folders and subfolders are used to organize
electronic files systematically on a hard drive. A
structural hierarchy is formed by nesting—the act
of placing one folder inside another and so on.
The folder at the top of a nested collection of
subfolders is called the root or parent folder.
The Computer CHAPTER 2 67
FIXED STORAGE
Fixed storage refers to a nonremovable chip or drive
permanently installed within the chassis of a computer or dig-
ital device. Fixed storage can usually be replaced or upgraded,
but it’s not designed to be quickly moved from one computer
to another. The most common type of fixed storage
medium is a hard-disk drive (abbreviated as HDD or hard
drive (see Figure 2.22). A hard drive is a mechanical device
that reads and writes data onto magnetically charged
spinning platters located within a sealed compartment.
Hard drives come in a variety of sizes and styles but can be
broadly classified into two main types: internal and
external (see Figure 2.23).
FIGURE 2.22
A vector illustration of a mechanical hard drive.
The Computer CHAPTER 2 69
FIGURE 2.23
An internal hard drive
is mounted inside the
chassis of a desktop
or laptop computer.
An external hard
drive is connected
to a computer via
an interface cable
(typically USB or
Firewire). External
drives sometimes
require a second
cable to provide
electrical power to the
device.
All things being equal, drive A is the better value, even though the total cost of
the unit is higher than drive B.
REMOVABLE STORAGE
The second category of computer storage is broadly referred to as removable
storage, and as the name suggests, it refers to a portable storage medium that’s
designed to easily move data from one computer to another. Removable storage
72 SECTION 1 Multimedia Foundations
Tech Talk
Optical Storage
Optical storage devices use a laser beam to read and encode data onto the
reflective inner surface of a hard plastic disk. The laser beam encodes binary
data as a series of pits and lands. Pits are microscopic depressions burned onto
the surface of the disc. Lands are the unaffected regions between the pits. The
transi- tion from a pit to a land or from a land to a pit is interpreted as a
one, while no change (pit to pit or land to land) indicates a zero. Pits are
recorded along a single, continuous spiral track from the inside of the disc to
the outer edge.
The Compact Disc While video laserdiscs hit the market in the late 1970s,
the Sony/Phillips audio compact disc (1982), or CD, was the first widely
adopted optical storage technology. An audio CD can hold 74 minutes of
uncompressed audio. As the first consumer digital audio format, the CD was
lauded for its size, superior sound quality, durability, and random accessing of
music tracks. Other uses for CD technology evolved, and in 1985 a variant of
the compact disc was released called the CD-ROM (compact disc read-only
memory). With a storage capacity of 700 MB, a CD-ROM can hold as much
data as 486 floppy disks. As a result, CD-ROMs quickly eclipsed floppies as
the primary distribution medium for software. When the Windows 95 operating
system was first introduced, it was made available to consumers on either 30
floppy disks or a single CD-ROM. For customers with a CD-ROM drive on
their computer, the installation was much easier and faster to complete. To
illustrate just how densely information is stored on a compact disc, if the
concentric data track was stretched into a straight line, it would be .5 microns
wide by 3.5 miles long (5.6 kilometers). A micron is one mil- lionth of a meter.
The Computer CHAPTER 2 73
The average width of human hair is approximately 100 microns.
74 SECTION 1 Multimedia Foundations
FIGURE 2.26
A red or blue laser beam is used to write and read data on an optical disc.
While certainly a step forward, the CD-ROM was limited because data was
permanently stored on the disc and could not be erased or rerecorded. The acro-
nym WORM (write once, read many) is used to describe this type of recordable
medium. A significant breakthrough occurred with the development of CD-R
and CD-RW consumer disks and drives. CD-R technology allows for the record-
ing of digital information onto a blank disc providing a convenient way of
backing up or archiving large amounts of data. Like CD-ROMs, CD-Rs can only
be written to once. CD-RW (ReWritable) disks work the same way, but they can
be erased and reused more than once.
DVD The second type of optical storage technology is the DVD, short for digital
versatile disc or digital videodisc. DVD technology was developed by Toshiba and
released as an international standard in 1995. Toshiba designed the DVD as
an alternative to distributing video on videotape—do you remember VCRs? A
stan- dard DVD holds 4.38 GB of data, more than six times the capacity of a
compact disc, and it can hold up to 133 minutes of standard definition video.
The DVD- ROM version of the technology, along with the subsequent
development of DVD±R/RW recordable drives, has made this an ideal medium
for high-density removable storage. DVDs are backwards compatible, meaning
that an audio CD will play in a DVD device. While employing the same basic
technology as the compact disc, DVDs can hold more data because the tracks
are narrower and closer together. If you could stretch out the data track of a
DVD into a straight
The Computer CHAPTER 2 75
FIGURE 2.27
While physically identical in size, the capacity of an optical disc differs widely depending on
what type of media is being used.
line, it would be 7.5 miles long (12 kilometers). Newer Dual Layer DVDs
hold around 8.5 GB of data, but the media is relatively expensive and is already
being made obsolete by Blu-Ray discs.
Red or Blue Laser Technologies Optical discs have traditionally used a red
laser beam to read and write data on a CD or DVD. Blu-Ray was developed
by a large consortium of manufacturers in the consumer electronics, personal
computer, and entertainment industries. The recording system uses a blue
laser beam, giving it a shorter wavelength than traditional red-laser DVDs,
and pro- ducing a narrower beam with a much higher degree of precision. Blu-
Ray discs can hold up to 25 GBs of data on a single-sided disc and 50 GBs of
data on a two-sided disc. The additional storage capacity permits
manufacturers to dis- tribute high definition (HD) video content and movies
to consumers, signifi- cantly enhancing the home theater viewing experience.
Blu-Ray is backwards compatible with all existing red-laser optical media.
Flash Memory
Also developed by Toshiba, flash memory is a solid-state chip technology used
in a variety of digital appliances. Originally designed as a removable storage
solution for portable devices like digital cameras and cell phones, flash memory
is compact, lightweight, and extremely fast. Unlike magnetic and optical stor-
age solutions, flash memory chips contain no moving parts, therefore providing
exceptionally fast seek times and high data transfer rates. As discussed earlier,
this is the same technology SSDs use.
There are two main types of flash memory products available: flash memory
cards and flash drives. Flash memory cards come in many shapes and sizes and
are used as a local source of removable storage for portable electronic devices
like cameras and digital voice recorders. Some of the more popular card formats
are branded as Compact Flash, SmartMedia, Secure Digital (SD), and Memory
Stick. A card reader is required in order for a computer or printer to read the
contents of a flash memory card.
76 SECTION 1 Multimedia Foundations
FIGURE 2.28
Each of the 3.5-inch
floppy disks pictured
above had a
capacity of 1.44 MB.
It would take 2845
of these disks to
equal the capacity
of a 4 GB SD
memory card
(roughly the size of a
thumbnail).
GREAT IDEAS
Cloud Storage
Cloud storage is a method of saving data to servers located off-site and managed by a third-party vendor. It has many o
best, a team of users attached to a local area network (LAN). People outside the network firewall are generally forbidden
Cloud storage addresses this limitation by allowing users to transmit and share documents across open networks. Have yo
The author of an uploaded file can invite others to edit the document online, granting them the status of a collaborator.
FIGURE 2.29
Cloud storage enables
multiple users to
collaborate on the
development of a
single document or
body of work. The
author of a shared
file or folder defines
the permissions,
a set of rules that
determines the level
of access granted
to collaborators. For
example, one person
can be limited to
read-only privileges,
allowing them to view
a document online;
while others are
permitted to download
a file or make
changes to it online.
The Computer CHAPTER 2 79
FIGURE 2.30
USB (1.0, 2.0, and 3.0) and IEEE 1394 (Firewire) are two of the most common protocols used
today for connecting an external device to a computer. For Mac users, Thunderbolt may
eventually replace Firewire 800 as manufactures provide connectivity support for the new
interface.
protocol (IEEE 1394b) in 2002 with a new connector design and a bandwidth
of 800 Mbps. The terms Firewire 400 and 400-iLink are used to describe
prod- ucts with the original IEEE 1394 interface. Firewire 800 and 800-iLink are
used to describe devices equipped with the advanced IEEE 1394b connector.
For a while, Apple provided both Firewire 400 and 800 ports on its computers.
Today, only the 800 version of the connector is supported. Fortunately, an
adapter cable or hub can be used to connect a Firewire 400 device to a
Firewire 800 connector.
The most recent contenders in the connector wars are USB 3.0 and Thunderbolt.
USB 3.0 builds on the older USB standards but increases data throughput to
3 Gbit/s. Thunderbolt, the most recent option, provides up to a 10 Gbit/s of
data throughput and allows users to connect an even broader assortment of
devices, including high-resolution video displays.
The personal computer is the core hardware technology used by multimedia producers
CHAPTER SUMMARY
to plan, create, and distribute creative projects. Like Morse code, computers use a
binary language system comprised of only two characters (zero and one). Each instance
of a zero or a one is called a bit. Combining eight zeros and ones into a single data
string forms a byte. The more complex the information, the more bits and bytes will be
needed to represent the information in digital form. The digital revolution has made it
possible for any form of human communication, including text, graphics, photographic
images, moving video, sound, and animation, to be translated into the language of
computers. Computer speed and performance comes down to how fast the various
components of a computer system and its peripheral hardware can process streams
of binary data. The term storage is used to describe the numerous devices capable
of recording digital information. The three main categories of computer storage are
fixed storage, removable storage, and network-attached storage.
This page intentionally left blank
CHAPTER 3
Planning and Design
71
Among all the types of paper documents, narrative fiction is one of Key Terms Audience AV Script C
the few that will not benefit from electronic organization. Almost Development Copyright Demograp
Intellectual Property Mockup
every reference book has an index, but novels don’t because there is Planning Postproduction Preprodu
no need to be able to look something up in a novel. Novels are Production Protype Screenplay Sc
linear.
Likewise, we’ll continue to watch most movies from start to finish. This
isn’t a technological judgment—it is an artistic one. Their linearity is
intrinsic to the storytelling process. New forms of interactive fiction are
being invented that take advantage of the electronic world, but linear
novels and movies will still be popular.
—Bill Gates, The Road Ahead (1995)
Chapter Highlights
This chapter examines:
The importance of planning and process in multimedia design
The use of conceptual frameworks for guiding design and production processes
The relationship between client and producer
The Previsualization tools used in the design process
The steps and substeps involved in multimedia design and usability testing
A ROAD MAP
A famous adage in advertising and marketing circles proclaims, “If you aim
at nothing, you will hit it every time!” This simple, yet profound cliché is a
call to the masses to be purposeful and intentional in planning for success. A
care- fully hatched plan provides direction, a road map for helping you get to
your intended destination. Without it, you will have no idea where you are
going or how to get there.
Imagine for a moment someone embarking on a 3,000-mile road trip across
the country without any planning or preparation. They didn’t pack any
clothes or provisions. They have no money or credit card. They didn’t even
bother to consult a map or install a GPS for guidance. Chances are they won’t
get very far
Multimedia Foundations. DOI: 10.1016/B978-0-240-81394-3.00003-3
© 2012 Elsevier Inc. All rights reserved.
72 SECTION 1 Multimedia Foundations
or have much fun along the way. While it’s difficult to fathom anyone like
this setting out on a journey so ill prepared, many people approach
multimedia design with a similar disdain for planning and process. Give a
person a video camera and he wants to start shooting right away. Tell another
person to create a website and she immediately sets off to build it. After all,
planning is so boring, and we want to get to the fun stuff as soon as possible.
In reality however, the time and effort you put into planning a multimedia
project will determine the degree to which the project, once completed, hits
its mark. It matters little what the plan looks like. What matters most is that
you have one; and that it sufficiently addresses the needs of the client, the
design team, and the users of the product. A good plan is one that is
thoughtfully crafted, thorough, and realistic. A plan is your road map and
serves as the foundational visionary framework undergirding the creative
enterprise.
GREAT IDEAS
Target Market and Target Audiences
A target market is a group of consumers that a company has strategically identified as
having potential to purchase and use its goods and services. Manufacturers seldom
produce a product that everyone wants or can afford. Because of this, they will go
through incredible lengths to find out as much as possible about consumers who are
the most likely candidates for using their products. The more they know about their
target market, the more successful they will be in designing and delivering effective
messages and product campaigns. When it comes to consumers of media products
such as television programs, newspapers, and websites, you’re more likely to hear the
target market referred to as the target audience. In effect, these are synonymous terms.
By understanding more about their target audience, media organizations can do a better
job of selling advertisements to companies looking to reach a specific target market.
FIGURE 3.1
“If you aim at nothing, you will hit it every time.” Identifying the target audience or user for
your multimedia product is a critical first step in the planning and design process.
Planning and Design CHAPTER 3 73
CREATIVITY
Experts in the field of creativity research generally agree that creativity is a
process that “involves the production of novel, useful products.” 1 This definition
suits us well, as the focus of this book is about making multimedia products
designed to communicate a message with deliberate thought and
intentionality. It’s not about art for art’s sake, not that anything is wrong with
that per se; but it is about the process of creating a meaningful communication
experience for the benefit of an audience. Such value will rarely emerge
without giving proper attention to planning and process.
Observe a young child at play and you will see that
creativity can flow spontaneously out of an active
imagination without any forethought or predeter-
mined goals (see Figure 3.2). Hand a child a draw-
ing tablet and some crayons, and in no time at all
the paper will be filled with lines, shapes, outlines,
and colors. The recognizability of such primitive
doodlings will depend entirely on the child’s age
and skill. Sometimes children will begin with an
idea, but oftentimes they are content to create on
the fly without any preconceived notions of where
their creativity and talent will lead them.
While there is a place for such unstructured
moments of creative expression, aspirations and
needs mature as we grow older. A carpenter will
not embark on building a house until the
blueprints
for the design and layout of the construction have FIGURE 3.2
been completed and approved. Likewise, a For the multimedia producer, creativity is a process that
profes- sional web designer must have a plan “involves the production of novel, useful products.”
before the actual coding and design work can
begin. A formal process of some sort is
necessary for translating any sophisticated idea or
concept into a multimedia product that people will
find useful. A project of any great size will require a
process with numerous steps from start to finish
and involve the work of many people with
specialized knowledge and skills.
The Three Ps
For many years, film and video producers have
utilized a popular process model known as the FIGURE 3.3
three Ps: preproduction, production, and The Three Ps Production Model.
postproduction (see Figure 3.3). The elegance and
simplicity of this
Mumford, Michael D. (2003). Where have we been, where are we going? Taking stock in creativity
1
74 SECTION 1 Multimedia Foundations
research. Creativity Research Journal, 15(2 & 3), pp. 107–120.
Planning and Design CHAPTER 3 75
FIGURE 3.4
Preproduction is the planning phase of the design process and includes a broad set of activities required for a
successful outcome.
PREPRODUCTION
Preproduction is the preparation phase of a project and involves many steps,
including: problem identification and needs assessment, idea generation, con-
cept development, research, audience analysis, scriptwriting, budgeting, sched-
uling, staffing the production team or crew, auditioning, scouting locations, etc.
(see Figure 3.4). Preproduction activities are meant to establish the scope, direc-
tion, timeframe, and cost of a project before spending one dollar on the actual
production. When done well, preproduction saves time and money during
sub- sequent phases of a project, which are almost always more expensive.
PRODUCTION
Production is the acquisition or design phase of a project (see Figure 3.5). In
a film or video project, this is the recording phase and includes: (1) all of the
work that goes into preparing for a shoot or recording session, such as build-
ing sets, hanging lights, miking instruments, applying makeup, setting up the
equipment, etc.; (2) the actual recording of shots, scenes, and tracks that will
be edited together to tell a story; and (3) striking the set, packing up the
equip- ment, and vacating the premises at the end of the day. When tape
begins to roll and the director hollers “ACTION!” you know that production
is in full swing. More generally, production involves the creation of any
media asset, including static graphics, photographs, time-based media, or
animations that are needed for inclusion in the final product.
76 SECTION 1 Multimedia Foundations
FIGURE 3.5
Production is the acquisition or design phase of a project where the individual pieces of a multimedia project are acquired
or produced.
GREAT IDEAS
Intellectual Property
While production often focuses on the creation of original content by the designers
and producers of a work, it can also include the legal acquisition of media resources
owned by others. Intellectual property is a term used to describe many types of
created works that are protected by copyright, trademark, or patent law. Just because
an image or piece of music can be easily downloaded from the Internet doesn’t mean
you can legally use it in your own work. Whether you intend to materially profit from
the use or not, or whether the work is intended for commercial or noncommercial use,
doesn’t always matter. Copyright law is complicated, and without the expert advice
of a lawyer who specializes in such matters, it is often difficult to know for sure what
permissible uses the law will, or will not, allow. When in doubt, it is always better to
err on the side of caution. The good news is that there are many avenues available to
the multimedia professional for acquiring photographs, music, clip art, stock video,
film footage, animations, and more from companies specializing in the sale of royalty-
free or licensed content.
It is also easy to find sites publicizing content that is free or in the public domain.
While some of these sites do indeed offer legitimate ways for securing legal rights to
use another person’s intellectual property, you need to be careful and discerning. You
also need to read the fine print. For example, a media resource may be allowed for
publication on the Internet but not in a program aired on television or in a derivative
work for sale, such as a DVD. As with anything, you need to choose your sources
carefully and dig deep to discover all the facts. In a court of law, you’re usually no less
culpable because you didn’t know any better, or because you were mistakenly led to
believe that it was okay to use content downloaded from an Internet site or scanned
from a book. For this reason, it is best to rely on companies with an established
reputation, who own the rights to the material you want to use, and who have the legal
authority to grant permissions for its use.
POSTPRODUCTION
So you’ve been out on location, or in a design suite, recording studio, or
sound stage, and all of the raw footage and material you need for your project
has been acquired or gathered. Now is the time for putting it all together.
Postproduction
Planning and Design CHAPTER 3 77
FIGURE 3.6
Postproduction
brings together the is the assembly phase of a project (see Figure 3.6). In traditional time-based
individual components workflows, postproduction is spent almost entirely on video and audio edit-
of a multimedia ing. However, it also includes the creation of titles, motion graphics, and
project into a unified video effects. A rough cut is the first complete edited copy of a program.
finished product.
The postproduction
Depending on a project’s complexity and the demands of the producers, a
workflows you adopt rough cut may go through several rounds of revisions before being approved.
will vary depending The approved edited master program, called a final cut, moves on for one final
on the project, for round of polishing, a process that may involve audio sweetening, color
example, whether correction, and other forms of fine tuning.
you are editing a film,
building a content
management system,
or constructing an
THE CLIENT AND PRODUCER
online photo gallery. Multimedia production can be a solo endeavor with a design team of just one
person, but more often than not it is a collaborative process involving many
people representing one of two parties: the client and the producer. In the
cor- porate sector, the client may be an individual but more likely is a
business or organization. The client is the content expert. Clients are the ones
most familiar with the mission, goals, and needs of the organization they
represent.
Likewise, the producer may be an individual as well. The market is full of free-
lance producers and directors who operate independently on a work-for-hire
basis. More likely, however, the producer is a creative agency or production
com- pany. The producer is the message design expert. Producers are the ones
most familiar with the process and tools for crafting an effective message and
deploy- ing it strategically to a target audience.
While some companies can turn to an in-house department or creative unit to
handle a communication problem, others don’t have this option; and even if
they did, they will sometimes choose to hire an out-of-house producer to get the
job done. Their corporate colleagues may be booked solid and unable to meet
the project deadline, or perhaps they just want the objectivity and fresh eyes
of an independent producer who is free of company bias and old ideas.
Questions will become more specific as the discussion unfolds and the
project moves forward. The initial client meeting will rarely lead to a
concrete idea or conceptual road map. Its purpose is purely informational, to
gather the facts and to give the design team enough information to begin
working on an idea, theme, or concept for the project.
GREAT IDEAS
Ethnicity/ Race
Gender
Income
Age
Employment Status
Religion
DEMOGRAPHICS
FIGURE 3.7
Demographic variables such as age, gender, and ethnicity are used to describe the shared traits of a u
2
Adams, A. (1995). The Camera. Boston: Bulfinch Press.
80 SECTION 1 Multimedia Foundations
and releasing the shutter. Under normal circumstances, she doesn’t need to
spend hours of preproduction time making sketches or thumbnails of shots
that will be taken tomorrow or next month. However, when producing some-
thing more complex, such as a film, website, or user interface, it pays to
refine and expand the concept on paper before proceeding to the production
phase of the project.
Since ideas are hatched in the mind, they must be put into a transferable form
before they can be communicated to others. Verbal brainstorming is a great way
to get the creative juices flowing during the formative stages of idea
generation and concept development. However, as an idea begins to take root
and grow, the methods used to develop it and communicate it to others need
to grow as well. Previsualization is a term that has been used for years to
describe the act of putting an idea or concept into a visual form that can be
better understood by the author of the idea and shared with others. Let’s look
briefly at six previ- sualization tools that have been widely used in
multimedia design: treatment, storyboard, script, wireframe, mockup, and
prototype.
GREAT IDEAS
Think Outside the Box
FIGURE 3.9
“Think outside the box” is a popular expression used by creative people to encourage the imagination to think beyond the ob
Planning and Design CHAPTER 3 81
TREATMENT
A treatment is a short narrative description of a project. It provides the reader
with a concisely written summary about a concept or story idea. A treatment
should include sufficient details to address the basic who, what, when, where,
and why types of questions. As appropriate, it should also include information
such as the project title, purpose, goals, audience, and genre. Treatments are
often used for pitching narrative film and video projects. A treatment is similar
in purpose to an executive summary or resume. It is a short-form device used
to cast a vision or sell an idea. The length of a treatment depends on the type of
project being proposed. A treatment for a feature-length screenplay for a motion
picture film might be four to five pages long, while one for a news story or
television commer- cial could be done in half a page or less. Treatments can be
used for describing any type of multimedia project, including web pages and
interface components.
STORYBOARD
A storyboard combines words and pictures together to communicate an idea (see
Figure 3.10). It is an organized panel of images and text used for previsualizing a
linear or nonlinear sequence of actions or steps. In the 1930s, Walt Disney pop-
ularized the storyboard format to assist artists in the production of animated
films. Today, storyboards are used routinely in multimedia production from
planning short-form motion picture projects such as television commercials and
“webisodes,” to interactive narratives used in web and interface design.
Resembling pages from a comic book, a storyboard is made up of a series of
boxlike frames depicting the visual portion of a story. Written copy is placed
adjacent to each frame (usually below it) to describe its contents. The text can be
purely descriptive or include spoken voiceover copy or dialogue. Before comput-
ers, storyboards were drawn by hand. Today, they can be generated any number
of ways using sophisticated previsualization software, page layout and design
programs, or online tools. In fact, a storyboard template is probably available
for your favorite word processing program.
SCRIPT
A script is a written narrative framework for organizing the visual and audio por-
tions of a multimedia presentation. Scripts are most often used for time-based
media projects and increasingly for interactive narratives.
AV Script Format
The two-column AV format is ideal for the producers of radio and television
commercials, music videos, promotional videos, education and training pro-
grams, documentaries, interactive narratives, and more (see Figure 3.11). It is
one of the most common and versatile script formats used in commercial pro-
duction today. This format features a divided page with two vertical columns.
The column on the left takes up about one-third of the width of the page and
includes information about the video portion of the program. The wider col-
umn on the right is used for audio information such as music cues, voiceover
82 SECTION 1 Multimedia Foundations
FIGURE 3.10
This storyboard was created to help visualize scenes from the screenplay Cupcake, written by Paul Castro. Filmmakers
and advertising agencies will often hire a storyboard artist to illustrate a script prior to shooting a film or television
commercial. (Courtesy of Paul Castro)
Planning and Design CHAPTER 3 83
FIGURE 3.11
This two-column AV
style script was used
in the production of
an educational video
program.
copy, or actor dialogue. The AV script format is used in both single-camera and
multicamera production settings.
Screenplay Format
The screenplay is the Hollywood studio format used in the production of
motion picture films and prescripted television comedy and drama pro-
grams (see Figure 3.13). This format features a single-column layout where
each page translates to roughly one-minute of on-screen action. Full-length
84 SECTION 1 Multimedia Foundations
FIGURE 3.12
The script in Figure
3.11 was quickly
transformed into this
single-column layout
using Final Draft AV, a
popular scriptwriting
program. Script
formats vary widely
depending on the
company or industry
you work for.
WIREFRAME
The design process for a new website often begins with a series of drawings or
sketches called wireframes (see Figures 3.14 and 3.16). A wireframe is a visual
representation of the layout or skeletal structure of a page. A designer will often
use a wireframe to experiment and pitch various ideas for the placement of
content and interface components that address the client’s needs and goals. First-
generation wireframes are often very simple and crude. In the beginning stages
of a project, the designer will often produce multiple wireframes for the
client to evaluate and choose from. Sometimes there’s a clear winner, and
sometimes the client doesn’t like any of them, forcing the design team back
to the drawing board for a second try. Another possibility is that the client
Planning and Design CHAPTER 3 85
FIGURE 3.13
This is a sample page
from the screenplay
August Rush. The
screenplay format is
an industry-standard
scripting convention
and is used widely
throughout the motion
picture and television
industries.
(Courtesy of Paul Castro).
might like a design element in one wireframe and ask to have it incorporated
with features in another. Wireframes are a great tool for honing in on the aes-
thetic tastes and sensibilities of the client without exhausting a lot of time and
money on the front end of a new project.
86 SECTION 1 Multimedia Foundations
FIGURE 3.14
A graphic designer
created this wireframe
for previsualizing the
structural layout of a
new web page and its
contents. The drawing
was created with a
computer software
program called
Balsamiq, a visual
mockup tool used
by professional web
designers.
(Courtesy of North Star
Marketing and the
Burlington/Alamance
County NC Convention and
Visitors Bureau).
MOCKUP
Once a wireframe has been approved, designers will often create a more
refined visual rendering known as a mockup (see Figure 3.15). A mockup is a
detailed actual-size version of the page, usually created by a graphic designer
with a program like Adobe Photoshop or Fireworks. It can include either actual
copy and artwork or temporary content such as image placeholders and dummy
text. A mockup also shows the color scheme and style elements of the page. The
purpose of the mockup is to provide the client and design team with a detailed
visual model of what the
Planning and Design CHAPTER 3 87
web page or interface component will look like when finished. As with
wireframes, multiple versions of a mockup will usually be shown to the client.
PROTOTYPE
Wireframes and mockups are great for experimenting with the visual look
and layout of a page or interface, however, they won’t help much when it
comes to
FIGURE 3.15
Once approved by
the client, the rough
wireframe sketch
shown in Figure 3.14
was transformed
into this detailed and
colorful mockup using
Adobe Photoshop. The
nonsensical Latin text
(commonly known
as filler or dummy
text) was used as a
temporary placeholder
to help the client and
design team more
fully envision what the
page will look like in
its final form.
(Courtesy of North Star
Marketing and the
Burlington/Alamance
County NC Convention and
Visitors Bureau ).
88 SECTION 1 Multimedia Foundations
FIGURE 3.16
For this project, the design team created wireframes and mockups for every section (or main page) of the website. Such
careful planning on the front end of a project helps ensure that time and money will not be wasted when the project
moves forward into production. (Courtesy of North Star Marketing and the Burlington/Alamance County NC Convention and Visitors Bureau).
testing the usability of a design. At some point before production begins, you
need to verify that a design is functional and will work as intended.
Otherwise, a great deal of time and money may be put into creating a project
that is techni- cally flawed or doomed from the start. For this you need a
prototype.
A prototype is a paper or working model of the project designed for
simulation and testing. As the name suggests, a paper prototype is built on
paper or cards and can be constructed using simple sketches or drawings. If
time and money allow, the quality of a paper prototype can be refined using
computer-generated artwork. Unfortunately, some functions, like scrolling or
drop-down menus, are difficult to mimic in a paper prototype, creating limits
on how much “real- life” testing can be done with a paper model. To
overcome such limits, pro- fessional designers will often use a prototyping
software application like Adobe Fireworks. Using application software, a
designer can build a computer-based prototype that more realistically
simulates the multimedia experience for the end-user. A prototype is like the
blueprint of a house. Once approved, the web development team and
programmers can begin the more time-intensive and costly step of actual
construction.
Planning and Design CHAPTER 3 89
Specify Requirements
Specifying requirements helps establish the scope of your project, or the
range of content and pages or screens. During this phase, your design team
will meet with the client to identify business needs and requirements,
existing problems, and potential users. You’ll also develop functional solutions
and decide what to include in the project and what to leave out.
Begin with a process known as benchmarking. Look at similar projects to see
what they contain and how they are organized. Do you want similar types of
screens? Different screens? Figure out the number and types of screens (or
web pages) you need. Give each a temporary name, then sort them into
logical groups and headings based on content. You can refine these categories
later when you work with real users by having them do card sorting, creating
your main navigation by letting them sort content into groups with
predetermined names or by naming groups themselves.
Analyze Users
Next, you will conduct research to define your target users by demographic char-
acteristics and/or their technology skills. You will explore when, where, and how
target users will use the interface. You’ll also learn about their previous experi-
ence with similar interfaces. The client is the best resource for this information
and can tell you who the primary and secondary users of the site will be.
However, don’t assume the client will tell you all you need to know about the
users. The client may have a fixed vision of what the site should include but may
have missed some obvious—or not so obvious—user concerns. Your job is to
ask the right questions.
If your client asks you to create a recycling website, first ask whom they
intend the site to serve. Will the site target people looking for a place to drop
off their used appliances? Or will it be aimed at local manufacturers trying to
understand recycling laws as they apply to small businesses? Or maybe it will
be aimed at grade-school teachers.
90 SECTION 1 Multimedia Foundations
Many interfaces, whether websites or other multimedia projects, will have two
types of audiences: a target audience (the audience you intend to reach) and
ancillary audiences (users who will use the interface for other reasons). Your
first concern is the target audience. This is the group of people your client most
wants to reach. If the site is designed to reach teachers, they are your primary
audience. If your client also wants to have information for local residents and
businesses, these are your secondary target audiences. Focus on designing for
your primary audience, but keep the site usable for secondary audiences too.
If you can meet with a few people who will use your interface, do it. If you can’t,
try to find people who are similar to real users. If that still doesn’t work, do your
best to think like real users. Start by answering the following questions about
your target audience:
■ Who are the users? Start with this, the most obvious question. How old
are they? How much do they use this technology, or a computer in
general? What do they use the technology for? Are they part of a
particular organiza- tion? And so on. If you’re designing a site for people
in their 50s, tiny fonts are a bad idea.
■ What tasks will they try to accomplish? What might they use this system or
site for? What are the most important of these tasks?
■ What conventions are my users used to? Look at other interfaces to develop your
design criteria (benchmarking). What do similar interfaces look like?
Does this one need to tie into a theme from a parent corporation? Don’t
copy another organization’s interface or site, but look for patterns.
■ What do they like about similar interfaces? What works on those interfaces in
terms of navigation, color, organization, and so on?
■ What do they find frustrating about similar interfaces?
■ What type of equipment and connections do my users have? If you’re building a
website or application, what types of computers and software do they
have? What size monitors and resolutions do they use? What types of
handheld devices will they use? How fast are their Internet connections?
If you’re designing a phone app, plan for a small interface. If you’re
designing a web- site that will be used by children in school or people
who work on small laptops, plan to design for a smaller screen.
■ What colors and images do they prefer?
Once you’ve divided content into sections, you need to choose how much space
to devote to various components, for instance, how many items to include in
a horizontal menu. The number of menu items will affect the size, and
therefore look, of your buttons or bar. Next, figure out what components you are
going to include on a given part of the interface. Consult with the client to see
what they want on it. List the various options they have. If you don’t, you’ll
find that they will often add components after you’ve worked hard on a
design that doesn’t have space for extra items. Put these items in a wireframe, a
line drawing show- ing the main components (main navigation, logo, footer,
etc.), and get their feedback. You can create a wireframe on the computer or
using pencil and paper.
It is often much easier to create a more detailed picture of what one page will
look like than to create a working screen or page. Then, when you find major
mistakes, you haven’t wasted valuable time. More importantly, you won’t be
so committed to the design (because of all of your hard work) that you refuse
to make necessary changes. The mockup is a more photorealistic version of a
page than a wireframe. Choose a medium you can work with easily. If you’re
good at Photoshop or a similar application, create a sample page using that
application. Or use a drag-and-drop mockup or wireframing application (such
as Mockflow and MockingBird) to easily make a visual of an interface,
including using your own pictures.
Create three or more options for your client to evaluate, and let them choose
an initial design. Then, have a few users look at your mockup. You can either
get general feedback on the design, or you can test your paper prototype,
conducting a mini-version of the testing you’ll do later on a working
prototype. You don’t need more than a few mocked-up pages or screens to
test, and a prototype test can save you lots of work later. After all, it would be
disappointing to go through the hard work of creating pop-up menus only to
learn that your users don’t like them. Usability testing, which we’ll cover later,
is best done early and often.
Next, turn your design into a working prototype with at least some functionality.
your design. No matter how painful, fix the problems now so your users will
be happy. And test again to be sure your changes work.
The preproduction process for a multimedia project includes all of the activities leading
CHAPTER SUMMARY
up to the client’s approval of the concept and initial design. Whether you’re producing
a video, website, or content management system, this process will likely begin with
an in-depth conversation between you and the client and all of the key
stakeholders in the project. Since creativity is largely a subjective process, a
multimedia message can be designed an infinite number of ways and still be
successful. Don’t waste your time trying to nail down what you think is the only right
way to go. Instead, spend your time wisely, listening and asking lots of questions. The
best ideas often emerge from a process that is thorough, thoughtful, inclusive, and
collaborative.
Planning and Design CHAPTER 3 93
This page intentionally left blank
SECTION 2
Multimedia Design
4. Visual Communication 95
5. Multimedia Page Design 127
6. Interface Design and Usability 153
7. Web Design 181
Marco Rullkoetter/Shutterstock.com
CHAPTER 4
This page intentionally left blank
Visual Communication
95
Design is the method of putting form and content together. Design, Key Terms Elements of Design
Dot Line Shape Form Color Textur
just as art, has multiple definitions; there is no single definition. Principles of Design
Design can be art. Design can be aesthetics. Design is so simple, Unity Alignment Proximity Similar
that’s why it is so complicated. Without aesthetic, design is either the Emphasis Contrast (Value) Color
Depth Proportion
humdrum repetition of familiar clichés or a wild scramble for novelty. Perceptual Forces Balance Contin
Without the aesthetic, the computer is but a mindless speed machine, Aesthetics Aspect Ratio Asymmetr
producing effects without substance, form without relevant content, or Axis Canvas
Composition Content and Form De
content without meaningful form. Design Dutch Tilt Eye Candy
— Paul Rand, American Author and Graphic Designer
Chapter Highlights
This chapter examines:
The process of visual communication
The role of content and form in visual space
Eight elements of design: space, dot, line, shape, form, texture, pattern, and color
Twelve principles of design affecting human perceptions of visual form in two dimensional space.
Unity: proximity, alignment, similarity, and repetition
Emphasis: contrast, color, depth and proportion
Perceptual Forces: balance, continuation, figure-ground, and psychological closure
VISUAL COMMUNICATION
Visual communication (sometimes referred to as vis-com) is an area of study that
investigates the transmission of ideas and information through visual forms and
symbols. On a deeper level, it also looks at the cognitive and affective processes
that affect the way we perceive (or sense) visual stimuli. Seeing is one thing and
perceiving is another. The former has to do with the objective realities of
sight, while the latter has to do with the transmission of culture and meaning.
Two people with equal vision can stand in front of a painting and physically
“see”
Multimedia Foundations. DOI: 10.1016/B978-0-240-81394-3.00004-5
© 2012 Elsevier Inc. All rights reserved.
96 SECTION 2 Multimedia Design
FIGURE 4.1
Seeing and perceiving are two different things. The patrons in this art gallery “see” the artwork on
the wall in much the same way (assuming equal vision). However, their perceptions of aesthetic
value and beauty are individually constructed. What one person likes, another may disdain.
Photographer: Adriano Castelli/Shutterstock.com
the same thing. The “seeing” part of the event will be pretty much the same
for both participants. However, each person’s experience, point of view,
knowledge, and aesthetic sensibilities, will shape his or her personal
judgments about the painting’s beauty or the artist’s ability. One of them may
like what he or she sees while the other may not (see Figure 4.1).
FIGURE 4.2
The Congressional lawmakers who penned the Bill of Rights provided the content for this
illustration of the First Amendment. A graphic designer created its form. Courtesy of the
Freedom Forum.
98 SECTION 2 Multimedia Design
FLASHBACK
The Medium Is the Message
The electric light is pure information. It is a medium without a message, as it were, unless it is used to spe
the “message” of any medium or technology is the change of scale or pace or pattern that it introduces in
Marshall McLuhan, Understanding Media: The Extensions of Man (1964)
Visual Communication CHAPTER 4 99
Aesthetics
We make perceptual judgments about visual beauty every day. We might
observe that a certain website looks amazing while another one appears drab and
unattract- ive. We turn up our nose at unsightly outfits in a store that grate
against our per- sonal taste, then rave when we find the one that’s just right for
us. Likewise, people have diverse, and often pronounced, sensibilities regarding
color. Some of us favor blue, while others prefer red. Think about the color and
style preferences that have
FIGURE 4.3
Our aesthetic
sensibilities vary
widely depending on
many factors including
context and place.
Which armchair is
more appropriate
for a fishing lodge
in the Appalachian
Mountains? Which
one would you rather
see in the office of a
lawyer?
100 SECTION 2 Multimedia Design
ELEMENTS OF DESIGN
Design is the strategic arrangement of visual elements within a two-dimen-
sional space to form a unified and holistic impression. Sometimes, the word
composition is used as an alternative way of describing the same process. For
example, in one situation it may be appropriate to say, “I like the design of your
web page or corporate logo,” while in another it makes more sense to say, “I
like the way you composed this shot” or “the composition is well-balanced.” In
both cases, I am referring to same thing: the intentional organization of a visual
space.
The elements of design are the fundamental building blocks of visual
content. In the same way that a house is constructed of essential components
like the foundation, floors, walls, windows, doors, and roof, graphic images
are formed using the elements of design. This section focuses on eight
elements of design that are common to all forms of visual communication:
space, dot, line, shape, form, texture, pattern, and color.
A popular cooking program pits two skilled chefs in a timed competition.
Each chef is given one hour to prepare a multicourse meal for a panel of
judges. At the beginning of each program, the host reveals the “secret ingre-
dient,” a specific food component such as corn, honey, fish, or peppers that
must be featured in each dish. Among other things, contestants are judged
on their ability to incorporate the secret ingredient into every course. They
will use other ingredients, but their goal is to make the secret ingredient
stand out as the star attraction. It may be helpful for you to think of the
elements of design as a master list of ingredients that you can use when
designing a graphic or composing an image. Typically, one or two of the ele-
ments in a design will contend for the starring role of “secret” ingredient.
Rarely do you see a design that will incorporate every element within a
single visual representation. Sometimes, the designer will choose to
accentuate the element of line. At other times the designer may want to
highlight form or color. By combining the elements of design, we can
create an endless array of variations.
Understanding the elements of design and their influence on human
perception will help you become a better visual communicator and media
consumer. It will also strengthen your ability to make informed judgments
about the overall quality and beauty of a visual design. Learning the language
of visual design will help you better defend and critique your own work as
well as that of oth- ers. Doing so will help you complete the sentence, “I like
this because . . . ” or “Photograph ‘A’ is better than Photograph ‘B’ because ”
Space
Like a traditional artist, a digital designer begins with a blank surface or
design space that comes to life as the creative process unfolds (see Figure 4.4).
In Adobe Photoshop and Illustrator, the document window or workspace is
referred to metaphorically as the canvas. Like an artist’s canvas, the digital
canvas is flat
102 SECTION 2 Multimedia Design
FIGURE 4.4
A visual
communicator’s job
is to fill empty space
with meaningful
content that
communicates a
message.
TWO-DIMENSIONAL SPACE
Digital workspaces only have two dimensions:
width and height. These dimensions are mea-
FIGURE 4.5 sured in pixels—rectangular or square picture
A camera
elements, or tiny blocks of color—rather than by using an absolute unit of
sees only
what the measurement, such as inches or centimeters (see Figure 4.6). As a result, the
photographer wants size is relative, because the actual appearance of the spatial area will vary
it to see. The design depending on the settings of the display device rendering the design on-
space is composed screen. We’ll talk more about pixels in Chapter 8, “Graphics.” The
within the frame of
workspace’s width is measured along its x-axis (horizontally), and its height
a camera’s
viewfinder and is is measured on its y-axis (vertically). When we refer to coordinates within
referred to as the the workspace, an x:y position, we set the upper left-hand corner as zero,
field of view. counting pixels from left to right and from top to bottom.
Visual Communication CHAPTER 4 103
FIGURE 4.6
This is a screen shot of the design space (or “canvas”) in Adobe Photoshop CS5. The blue
gridlines are a designer’s tool used for placing and aligning visual elements within the frame.
Digital spaces are rectangular, so they conform to the boxlike frames of elec-
tronic displays and projectors. The term aspect ratio refers to the relationship
between the width and height of a design space. A 4:3 space is 4 units wide by 3
units high. A 16:9 space is 16 units wide by 9 units high. Although the physical
dimensions of a display monitor or viewfinder can vary from small to large, as
long as the aspect ratios are compatible, images will stay true to their original
proportions. Using compatible aspect ratios for output or presentation prevents
unwanted distortion.
Digital spaces are physically limited to width and height, but people have the
ability to perceive the third dimension of depth when viewing two-
dimensional representations. Depth in two-dimensional space is only a
perceptual illusion. Through skilled control of lighting, perspective, shading,
texture, color, contrast, relative speed (for media with movement), etc., visual
designs can be made to look remarkably three-dimensional. Depth is
represented by the z-axis, an imaginary sight vector that extends away from
the viewer and into the center of the design. Our ability to perceive depth
depends on our ability to make sense of the relative position of background,
mid-ground, and foreground elements along the z-axis. As creatures of three-
dimensional space, we tend to favor designs that promote a sense of depth. A
design or image that is lacking in depth is said to be flat.
104 SECTION 2 Multimedia Design
FIGURE 4.7
White space (also
called negative space)
doesn’t have to be
white.
Tech Talk
The process has become more sophisticated over the years, but we still use
halftone printing today. In process printing, printers make screens for each
primary color channel: cyan, magenta, yellow, and black, or CMYK. During
printing, each primary color is applied separately to control the density of its
ink droplets. Under the microscope, color dots appear as clusters of four dots
of varying sizes, each representing the intensity of the color pigment needed
to make the final color. And we use similar principles to form electronic
images. Television and computer images are made of millions of pixels—each
one using various intensities of red, green, and blue to create a tiny square
of color. In printing, we refer to the tiny droplets of ink on a page as dots. In
digital imaging, we refer to the square points of light on a display screen or
projection as pixels.
Line
A line is the visual connector between two points in space. Lines can be real
or implied. For example, a white line painted on the edge of a road is real.
We know that the line did not just appear. It was intentionally painted to
warn driv- ers not to run off the road. Lines can also be implied by natural
alignment or the
Visual Communication CHAPTER 4 107
FIGURE 4.10
Lines can be
horizontal, vertical,
curved, or diagonal,
and can be combined
in an infinite number
of ways. The visual
world is full of them
in a myriad array of
shapes, sizes, and
colors. Take a look
around you. What kind
of lines do you see?
Can you describe
them?
STRAIGHT LINES
Straight lines can be static (horizontal or vertical) or dynamic (diagonal). Hori-
zontal lines often communicate a sense of peace and calm. Perhaps this is
tied to the fact that we lie down to rest and stand up to move and walk
around. Placing a flat object on a level surface such as a desk or countertop,
we would expect it not to move. The laws of physics should ensure a period
of stasis. We perceive horizontal lines in much the same way. They are a
natural and reli- able reference point that can reinforce a sense of balance
and stability within
108 SECTION 2 Multimedia Design
FIGURE 4.11
The parallel tracks
appear to converge
along the z-axis,
contributing to an
illusion of depth in
this two-dimensional
photograph.
CURVED LINES
Curved lines can create a sense of peace and tranquility in a design. The smooth
edges of a curved element provide a sense of flow and directionality and are easy
for the eye to follow without abrupt stops or interruptions. Car bodies are
typi- cally curved to improve the aerodynamic performance of the vehicle, but
curves also improve the appearance of a car, making it look sleek and stylish.
Curves accentuate the shape and form of other elements and are among the
most com- mon lines found in nature.
Visual Communication CHAPTER 4 109
Tech Talk
Shape
A shape is a two-dimensional element formed by the
enclosure of dots and lines (see Figure 4.14). We perceive
shapes as flat objects without depth (either real or implied).
A shape can be as simple as a circle or as complex and
intricately conceived as a snowflake. The basic geometric shapes
are circles, triangles, and squares. When we combine them in
various ways, they can form virtually any other shape
imaginable. They are called geometric shapes because they can
FIGURE 4.13 be rendered mathematically using for- mal rules of
In this photo, an S-curve is used to construction.
gently guide the viewer’s eye along a
linear path to the main subject. Organic shapes are so called because they resemble objects in
the natural world. Organic shapes have an imperfect, soft,
110 SECTION 2 Multimedia Design
FIGURE 4.14
Shapes often connote a sense of the familiar. Like lines, they can be combined in an infinite number of ways to form new,
more complex shapes.
Form
Form adds the dimension of depth to shape. Form is three-dimensional and
connects us more fully to the way we see objects in the natural world (see
Figure 4.15). In terms of geometric elements, it may be helpful to compare
the two-dimensional shapes of a circle and square to their three-dimensional
counterparts, the sphere and the cube. In order to show depth within a two-
dimensional design space, a designer or photographer must learn to manipu-
late lighting, shading, color, and contrast within the frame. In film, a single
Visual Communication CHAPTER 4 111
FIGURE 4.15
A 3D modeling
program was used to
create these visual
forms. Form adds
a sense of depth, a
perceptual illusion
that can be produced
by manipulating light,
shadow, color, and
other elements within
2D space.
LIGHTING
As we’ve already established, lighting affects forms. Good designers carefully
plan their lighting when they take photographs, capture video, and create graph-
ics. We describe light as having both color temperature and hardness. Color
temperatures, described in degrees Kelvin, can be low on the scale and warm
(orange or reddish) or high on the scale and cool (bluish), giving images differ-
ent qualities. Hard lighting from distant or narrow light sources casts shadows
and will give your image an illusion of volume and emphasize textures. If you
angle a light to the side of a texture (called short lighting in portrait
photography), small shadows will emphasize the lines in a face or the grains of
sand of a dune. Soft lighting from broad (frontally placed), multiple, close, or
diffused (scat- tered) light sources—even lighting diffused by clouds or from
backlighting—can soften an image, making an older subject appear a bit younger
or creating an ethereal quality.
Low light, universal light, and light falloff also affect an image. In low light, a
still-image camera has to keep its shutter open longer, and objects in motion
112 SECTION 2 Multimedia Design
can become blurry; in video, the image becomes grainy. And when you work
with images and graphics in editing applications such as Photoshop, using
universal lighting will make light-dependent effects (drop shadows and bev-
eling, for instance) more cohesive. Universal lighting makes the illusion of
light consistent, so light appears to come from only one angle. Finally, you
can light a scene to emphasize selective parts of a composition, taking advan-
tage of light falloff, or the lower illumination of objects farther from your
light source.
Texture
Texture is the surface attribute of a visual object that evokes a sense of tactile
interaction, and it can be implied in images. We perceive texture with the sense
of touch. Some surfaces feel smooth whereas others are coarse. Some are
wet, and others are dry. Some are soft, others are hard. Texture can affect us on a
mul- tisensory level, evoking in us a sense of touch, smell, and even taste. A
digital image can only imply a sensory response, but the power of suggestion is
real. As with shape, the visual element of texture can stir up memories and
associations in the user. The sight of grill marks on a hamburger can remind
us of the sound of sizzling fat and stir up thoughts of familiar smells and flavors.
The grill marks attest not only to how the meat was cooked, but also may
bring to mind posi- tive recollections of summer cookouts and backyard
parties. Of course, those who don’t eat meat may respond entirely differently
to such a visual prompt. Knowing your audience and its associations is
important, especially when your audience is intercultural.
Looking at a familiar texture on screen, you can easily imagine what it might
feel like. Adding sound to the experience makes the connection even
stronger. What if you could hear the burger sizzle? Just imagine the grinding
sound of sandpaper rubbing up and down the edge of a board, and the
relative graini- ness of the surface becomes all the more apparent. What
about the sound of feet walking on a gravel road? Does it change depending
on the size of the gravel?
Texture ignites the imagination, allowing us to interact with a design at a
deeper cognitive level. For example, the texture of a person’s skin can tell us
a lot about an individual. We associate soft, smooth skin with the touch of a
baby’s face. Likewise, we expect older people to have dry, cracked, or wrinkled
skin that is weathered by age and the elements of life. Rust and peeling paint
also speak to us about age and the passing of time. And the designs of many
websites and applications play to our expectations of current technology as
smooth and sleek.
Texture also serves to break up visual monotony. It adds visual depth, since
most textures rise above the surface of an object. When properly incorpo-
rated into a design, texture can enhance the interest and perceived realism of
a visual work.
Visual Communication CHAPTER 4 113
Pattern
Pattern is the reoccurrence of a visual element within a design space.
Clothing, furniture, and wallpaper are often identified by the characteristic
pattern they employ (checkerboard, herringbone, polka dot, paisley, plaid, etc.).
As with tex- ture, pattern can add visual interest to an otherwise plain and
monochromatic object. Like shape, pattern can be geometric or organic. The
Spirograph toy, invented in the 1960s, is a popular plaything that enables
kids to draw intricate geometric patterns of lines, shapes, and colors. The
device, made up of a drawing frame and design templates with gears and
tracks to control movement, allows the child to combine curves, circles,
loops, and lines to form highly complex patterns and designs. The long-term
popularity of this toy testifies to our innate fascination with repetition and
pattern.
1 2
3 4
FIGURE 4.16
1) Lighting, 2) texture, 3) pattern, and 4) color can be intentionally highlighted to achieve a
desired effect or mood. As with all visual elements, they can be mixed. The umbrella shot
equally combines the use of pattern and repetition.
114 SECTION 2 Multimedia Design
Organic patterns are found in nature, such as a flock of geese flying in formation,
the woven matrix of a spider web, or an aerial view of a pumpkin patch at
har- vest time. Patterns can be uniform and predictable as in the repetition of
colors and shapes in the American flag. The random or natural scattering of
similar ele- ments within a shared space can also lead to interesting patterns,
such as hiero- glyphics on a cave wall or shots of sunbathers strewn along a
crowded beach.
Color
Color has three dimensions: hue, saturation, and brightness. Hue is the color
shade of an object as a single point on the color spectrum. We refer to colors
most often by their hue (red, green, blue, etc.). Saturation is the strength or
purity of a color. The red in a stop sign is a highly saturated hue, whereas
pink is a desaturated version of the same color. A completely desaturated
color contains only variations of white and black as measured along the
grayscale. Brightness (also called value) is the relative lightness or darkness of
a color. Brightness can be thought of as the dimmer control on a light switch
that raises or lowers the level of light in a room.
FIGURE 4.17
Boy or girl? What role
does color play in
your decision?
Visual Communication CHAPTER 4 115
We can use color to set tone and mood, to elicit instant associations, to attract
attention, and even to help users remember things. In the beginning of the
movie Catch Me If You Can, Leonardo DiCaprio’s character is in control of his
life, and the palette is subdued. The colors become vivid as he loses control
and takes bigger chances. You might buy a beach towel with an orange and red
design because the colors bring to mind oceanside sunsets and stir up related
emotions. And the orange can generate positive emotions on its own, without
bringing a past trip to mind. Or you might paint a baby’s room a pale, desatu-
rated green to create a soothing environment.
Colors also have cultural associations. What sex is a baby dressed in blue?
(see Figure 4.17) Does a website with a pink background target men or women?
These are purely learned associations: in Western nations and prior to the end of
World War II, pink and blue were often used interchangeably across genders.
What does red mean to you? That depends on where you’re from. In the
United States, red is associated with danger and passion, among other things,
whereas in China, it has a number of other associations, such as luck. If you
want to attract attention to an element on a page, what color would you use?
Brown would not be your first choice! And psychologists have tied color to
memory, too. If you have used a coloring book to study anatomy, color may
have helped you pass your exams.
Unity
You have heard the saying, “the whole is greater than the sum of its parts.” In
visual theory, this means that the viewer should be able to grasp the essence
of the “big picture” without being distracted by the individual elements of a
design. This is the principle of unity, which is achieved when the visual
subcomponents of a design coalesce to form a holistic impression. Unity means
that the constituent parts of a work reside together in harmony. In a good design,
each visual element or group of elements should contribute to the whole
116 SECTION 2 Multimedia Design
without competing against other elements or distracting the viewer from the
primary point of interest.
Visual Communication CHAPTER 4 117
PROXIMITY
The law of proximity states that objects are more likely to be perceived as related
when they are positioned close together. Think for a moment about how Google
returns search results (see Figure 4.18). Google gives you a series of search snip-
pets, each with a linked web page title, URL (web address), and excerpt or
page description. Evenly distributed spaces separate the listings, allowing you to
iden- tify each separate snippet: all of its material is grouped together.
FIGURE 4.18
Observe how proximity is used to group related elements in the layout of a Google search
results page. Perceptual grouping makes it easier for the viewer to process and consume visual
information while enhancing the overall unity of a design.
Source: google.com
118 SECTION 2 Multimedia Design
ALIGNMENT
The principle of alignment encourages designers to position objects that belong
together along a common edge or implied line. Visual objects—text,
graphics, etc.—are normally aligned along their outer edges (left, right, top,
or bottom) or centers. However, any line will do, as long as the placement is
thoughtful and consistent. When composing text, we use the terms left, right,
and center justified to indicate which edge to line up against. To enhance
readability, we almost always left justify paragraph text or body copy, providing
readers with a consistent starting point at the beginning of each line.
Depending on the designer’s preferences and the nature of the content, the right
edge can remain ragged (uneven) or each edge can be justified (both flush
left and flush right). Leaving the right side ragged preserves the uniformity of
the word spacing and gives readers visual reference points in longer blocks
of text, which enhance readability. Justifying both sides of a paragraph adds
spaces between words to expand each line to its outer limits; it affects
readability but gives a design a clean look.
Within a single design frame or multipage layout, alignments should be
consis- tent. Mixing them is normally not recommended. If you choose to left
justify the body copy on one page, be sure to apply the same formatting to
every page in the site. Alignment can also be used to unify different, yet
related, visual elements. For example, left justifying a photo caption along
the left side of the image it refers to ties both elements together along a common
vertical edge. This lets the viewer know that they belong together. Be careful
not to have too many implied lines: a handful of well-chosen lines enhances a
design, but too many make it look cluttered.
SIMILARITY
The law of similarity states that the brain will perceive visual objects as
belonging together when their style attributes are similar and uniform.
For example, the unity of
this paragraph is diminished by the ARBITRARY
ANd random mixing of dissimilar stYLEs.
Here, a lack of similarity results in poor organization of the content and
dimin- ished readability of the text. The more objects are alike, the more
likely they will stick together naturally as a group. The more dissimilar they
are, the more likely they will resist grouping and pull apart. The law of
similarity does not
Visual Communication CHAPTER 4 119
REPETITION
Repetition is related to similarity and suggests that repeating visual elements
such as lines, colors, shapes, and patterns help strengthen the overall unity of a
design. In web design, repetition can be used to bring harmony and
consistency to the look of a multipage website. Repeating important symbols
or components such as a banner graphic or menu bar at the top of each page
enhances usability and brings visual synergy to the site. Similarly, it’s common
to see visual elements or scenes from a movie repeated in the jacket design of
a DVD and on the movie poster and website promoting the film. Repeating
elements can help a company to preserve its visual brand identification across
multiple communication chan- nels as it promotes its products and services to
consumers.
Repetition occurs frequently in nature, and photographers enjoy capturing it
on film. Repetition produces pattern, and patterns are intriguing to the eye.
Think about how stable and fascinating the pattern of a honeycomb is.
There’s some- thing innately interesting about its uniform structure and
appearance that cap- tures our interest and provides a natural sense of awe
and wonder.
FIGURE 4.19
The color pencils in the first vector illustration are haphazardly arranged. The second illustration
features a more purposeful arrangement, applying the laws of proximity and alignment. How
is your perception of each version influenced by the use of proximity, alignment, similarity, and
repetition?
120 SECTION 2 Multimedia Design
FIGURE 4.20
How are proximity, alignment, similarity, and repetition used in the layout of this web page design template?
Emphasis
The principle of emphasis suggests that a good design must have a primary focal
point or center of interest. In a newspaper, the headline type is set to a very large
font size in order to draw the attention of the reader to what the editor
believes is the most important story of the day.
Emphasis can be used to quickly guide the viewer’s attention to the main subject
or message in a communication exchange. Large bold headings on a web
page help you navigate quickly to the item or information you are most
interested in. Google’s homepage clearly emphasizes the text entry window
to make it easy for users to enter keywords in a search. Just imagine if all of
the text in a newspaper or on a website were exactly the same size, style, and
color. The usability of the content would suffer tremendously, and users would
quickly get frustrated and lose interest. Since our eyes are naturally drawn to
larger objects in a design, varying the size of spatial elements is a common
way to connote emphasis. Although there are many ways to create emphasis in a
design, we will look briefly at contrast, color, depth, and proportion (see
Figure 4.21).
CONTRAST (VALUE)
The term value describes the range of light and dark portions in an image or
design. In photography, film, and television, value is usually expressed as
con- trast. The strongest visible contrast is represented by the difference
between black and white. One of the reasons that books are traditionally
produced with black text on a white page is because this method creates the best
contrast for reading. The text “pops” out against the lighter background and
makes reading more pleasant and efficient. The text is the main event, and
the background should not present a distraction by competing for the reader’s
attention.
Visual Communication CHAPTER 4 121
The number of colors or gradient steps in a composition that fall between the
bipolar extremes of black and white is called tonal range. The greater the
tonal range, the greater the contrast, and the more interesting a composition
will gen- erally appear. The eye is attracted first to the areas of a composition
with highest contrast (or high-key), regardless of whether they contain color.
Increasing contrast is one of the easiest ways to emphasize the main subject or
visual object. In photography, contrast impacts image detail, clarity, and mood.
A low-key image contains mostly dark tones or color levels and communicates
a serious, somber, or reflective mood. Such images can also ignite a sense of
foreboding, mystery, horror, and fright. Photographic film was touted for years
because of its ability to reproduce quality low-key images. Most narrative film-
makers prefer shooting on film for this very reason. Achieving low-key tonality
is necessary in order to create realistic and compelling moments of heightened
visual drama. Digital cameras have come a long way, though some professionals
still prefer film’s ability to handle low-contrast lighting situations. A high-key
image is characterized by bright tones with very few dark areas, and a mid-key
image falls in between. The higher the overall contrast in an image, the brighter
and more cheerful the overall impression will be. In photography, film, and
videography, pushing contrast too far in either direction will lead to an unusable
image. This occurs when an image is under- or overexposed due to poor lighting
in the scene or on the main subject.
FIGURE 4.21
Contrast, color, depth,
and proportion work to
emphasize elements in
a design.
122 SECTION 2 Multimedia Design
COLOR
Color is a powerful tool for enhancing contrast in visual design space, and color
contrast has been used particularly well in the advertising industry. Perhaps
you’ve seen a print or television ad in which a single colored object (like a
hat or a can of soda) stood out against the backdrop of a black and white or
mono- chromatic setting. A selective splash of color can immediately grab
our atten- tion, providing a sense of focus and direction for the viewer.
Colors are classified as warm, cool, or neutral, depending on which portion
of the color spectrum they fall into. Warm colors reside near the orange
area of the spectrum and include shades of red and orange, as well as warm
greens. The human eye is attracted to the warm color regions of a design
first. We tend to perceive warm colors as breaking free from the background
and advancing toward us from a printed page or screen. A designer can use
this effect to bring attention to an object by increasing the figure-ground
contrast. But be careful! Too much warm color in a design can be visually
overwhelming. Cool colors such as violets, blues, and cool greens seem to
recede away from us into the background, appearing distant and detached.
Cool colors are calm, soothing, and placid, like the blue waters of a still
mountain lake. A design can accommodate large areas of cool colors without
visually overwhelming the user. Neutral colors are achromatic (effectively
lack hues). We’ll get more into color in Chapter 8, “Graphics,” but white,
black, and achromatic grays work well with warm and cool colors. Another
caution: many grays are not achromatic, and their hint of hue limits how
effectively they can be used.
DEPTH
Depth is related to the principle of figure-ground, or what we perceive in the
foreground or background, which we’ll discuss later. It is a powerful tool for
achieving emphasis in a design. In photography, film, and videography, the
term depth of field describes the portion of the z-axis that viewers perceive as
being in focus at any one time. When this distance is small, the image is said
to have a shallow depth of field. When the distance is large, we refer to depth
of field as being great or expanded. You can emphasize the main subject by
keeping it in focus (sharp and clear) and the background clutter or dead space
out of focus. In graphic design, drop shadows, lighting, perspective, and
other visual effects can be applied to a foreground image to simulate the
appearance of depth.
Visual Communication CHAPTER 4 123
PROPORTION
Proportion is the scale of an object relative to other elements within a composi-
tion. Our perception of an object’s size is related to the size and position of
other objects within the field of view. Decreasing the size of an object makes
it recede into the distance, creating the perception that it is getting farther and
farther away. Increasing the size of an object elevates its visual status, making it
appear closer to the viewer. The location of an object also affects our perception
of size. Objects placed near the top of the screen appear smaller and farther away
than objects placed near the bottom. Including familiar objects of a known
size within a design can help a viewer to make accurate perceptual judgments
about the actual, or relative, size of other elements nearby.
Perceptual Forces
When we look at a graphic representation within a frame (still picture or
moving image), our brains are constantly processing the relative push and pull
of percep- tual field forces within the visual space. For example, regardless of
whether we view a photograph when it is lying flat on a table, hanging on a
wall, or even rotated a quarter turn, we tend to perceive the top of the image
as up and the bottom of the image as down. The bottom of the frame
simulates the natural gravitational field of our human experience. But unlike
gravity, which acts upon us from only one direction, perceptual field forces can
tug at visual matter from any direction.
When a visual element is positioned too close to the edge of the frame, the
attraction between the object and the side of the frame increases. The effect is
similar to bringing the opposite poles of two magnets within close range. As the
magnets get closer, the attraction increases, until they eventually snap together.
White space can provide necessary breathing room around visual objects and
will help counteract the natural pull of the frame. When we discuss the percep-
tual forces humans experience, we talk about the principles of balance, continu-
ation, figure-ground, and psychological closure.
BALANCE
A balanced composition is achieved when the visual weight of objects is equally
dis- persed within the frame, producing a perceived state of equilibrium. On
occasion, the designer may want to purposely destabilize an image for effect,
but generally our goal is to achieve balance in design. The size, color, and
position of graphical elements affect our sense of balance or instability within a
composition. Like a set of scales, the frame rests on an imaginary fulcrum at
the center of its baseline. To the viewer, some objects will naturally appear
heavier or lighter than others. Typi- cally, we perceive large or dark-colored
objects as heavier than small or light-colored ones. If the designer places a large
dark-colored sphere in the upper left-hand corner of the frame, then an object, or
multiple objects, of equal combined mass will need to be positioned in the lower
right hand corner to achieve balance. Compositional balance can be achieved
using either a symmetrical or asymmetrical approach (see Figure 4.22).
124 SECTION 2 Multimedia Design
Symmetrical Balance
In a symmetrical composition, objects of similar shape, color, and size are weighted
equally on opposite sides of the frame. This can be accomplished by
centering the subject, or multiple visual elements, along the vertical or
horizontal dividing line of the frame, with the weight of the elements evenly
distributed in each half. Symmetrical balance is analogous to a seesaw whose
pivot point or fulcrum is centered, and children of equal weight are
positioned on opposite ends. This approach is also called formal balance
because it leads to designs that are per- ceived as tranquil, elegant,
traditional, and conservative. Symmetrical balance can lead to a somewhat
predictable and less-creative visual design.
Asymmetrical Balance
With asymmetrical or informal balance, equilibrium is established with objects
of differing size, color, and tone. This approach is analogous to a seesaw whose
ful- crum is located in an off-center position. In this configuration, placing two
chil- dren of equal size on opposite ends of the seesaw will not lead to
balance. More weight must be added to one side in order to compensate for
the new pivot point. Asymmetrical, or informal, compositions are much more
interesting to design and compose, and can appear more visually interesting
and dynamic.
FIGURE 4.22
Symmetrical balance
(top) and asymmetrical
balance (bottom)
work in different
ways to create
visual equilibrium.
Visual Communication CHAPTER 4 125
Tech Talk
FIGURE 4.23
Using the rule of thirds as a guide, how would you critique the composition of these two photographs?
126 SECTION 2 Multimedia Design
CONTINUATION
The law of continuation suggests that our brains tend to process what we see
as continuing along lines that are predictable and free of obstacles, and that
don’t abruptly change direction. So in a design with overlapping curves, our
brains interpret each vector (line) as a smooth curve; we don’t see each as
changing direction at their intersection. More precisely, vectors are
directional forces within the frame that guide the eye from one point to
another. Three types of vectors occur most often in visual space: graphic
vectors, index vectors, and motion vectors (see Figure 4.24).
Graphic vectors are created by strategically placing stationary line elements within
the frame. The lines can be real or implied through the placement of other visual
elements. They focus the viewer’s gaze in a specific direction or guide them
along a visual path. In photography, these pathways are called leading lines
because they “lead” the eyes from point to point or directly to the primary
focal point of the image. A barbed wire fence stretching across a landscape, the
line created by the water’s edge at the beach, and a row of people waiting for
a movie are all examples of graphic vectors.
Index vectors are created by placing objects that conspicuously point in a
specific direction. A crowd gazing upward, a road sign with a large arrow
pointing to the right, and a hitchhiker with his hand extended are visual examples
of index vectors.
Motion vectors are created by the real or apparent movement of subjects
within the frame. Motion connotes a sense of directionality. We can usually tell
which direction a person or object is moving by observing which way it is
facing. Although best demonstrated with time-based media, motion vectors
can be implied in a still image or graphic. The addition of a motion blur or
trailing edge can intensify the perception of speed and directionality.
FIGURE-GROUND
In Gestalt psychology an element that appears in the foreground of our
percep- tual field is called a figure, whereas everything behind it is the ground.
In nature,
FIGURE 4.24
Visual examples of graphic, index, and motion vectors (L-R). Continuation works to guide our eyes from one point to
another within the frame.
Visual Communication CHAPTER 4 127
FIGURE 4.25
Some optical illusions
work by playing with
the figure-ground
relationship. What do
you see in this image?
Relax your focus and
look again. How about
now?
PSYCHOLOGICAL CLOSURE
The final perceptual force we will look at is psychological closure. One of the
most powerful abilities we have is to mentally complete a visual pattern or
impression when only partial information is provided. Psychological closure is
the human equivalent of connecting the dots or filling in the gaps. For this
reason, objects within the frame can extend past the boundary of the frame. As
long as enough visual information is provided, an individual can complete a
picture in her head, thus maintaining a stable perceptual experience (see Figure
4.26).
The principle of psychological closure is used in visual design all the time.
Literal interpretations of words and symbols can be replaced with partial
impressions or abstract variations of larger elements. With psychological
closure, the brain seeks to create meaningful order out of visual chaos. As long
as enough visual cues are provided, the brain will kick in to form a complete
mental impression. The extra measure of sensory activity can give us a deeper
and more satisfying experience because we are participating and cognitively
engaging with the visual stimuli more.
128 SECTION 2 Multimedia Design
FIGURE 4.26
Kanizsa’s Triangle
illustrates the principle
of closure to suggest
a second triangle
overlaying the first. What we see... What we perceive...
Content is still king, but if we deliver content without good design, users will not
CHAPTER SUMMARY
know where to focus, will not make the connections we intend, will not be
engaged, and may even misunderstand the message. The elements and principles of
design are a designer’s foundation. They give you the tools to understand aesthetics
and preferences, and to both make and defend sound choices. Rarely do we work solely
for our own pleasure. We often work in teams, and we work for clients who hire us for
our expertise. Occasionally, others make suggestions that go against our better
judgment. Being able to explain why an idea won’t work well may make the
difference between creating an outstanding graphic or video and getting stuck creating
something that will fall flat with its audience.
Ultimately, we create for our users, the people who will visit our websites, see our
photos, watch our podcasts, and use our kiosks. Designers strive to make products
usable and engaging. Good design makes users’ experiences positive and rewarding.
Bad design serves no one.
A foundation in design elements and principles inspires not only good practice but also
creativity. Return to the foundation when you feel that you are in a creative rut. Do you
need to add energy to a graphic? Perhaps you can add dynamic lines. Are your
action photographs missing the action? A canted angle might help. Does your video
need to evoke other senses? Consider lighting your subject so it has more texture,
and add appropriate sounds. Are you at a loss for how to create a new logo? The
principle of clo- sure may suggest a creative solution. The combinations are limitless,
but be judicious: don’t emphasize too much. If you do, you might emphasize nothing at
all, leaving your users to wonder what your secret ingredient was supposed to be.
CHAPTER 5
Multimedia Page Design
127
The more personal computer displays become like lightweight books, Key Terms Body Copy Border
the more people are going to feel comfortable reading from them. Box Model Cell Chunking Column
Dynamic Page F-Layout Fixed Lay
A PC that you can open up and physically handle easily, that has the Fluid Layout Golden Ratio Grid S
right kind of battery and display, will give the same visual impression Gutenberg Diagram Headings
as a page. Headline Layout Margin Modular
Multicolumn Grid Padding
—John Warnock, Inventor of PostScript and the Founder of
Page
Adobe Systems (1994) Page Template Row
Sidebar
Single-column Grid Splash Page
Chapter Highlights Static Page Style Sheet Table
This chapter examines: Typographic Grid Visual Hierarch
Strategies for the effective placement of visual content within the page
The influence of visual hierarchy on viewing behavior
Use of the grid system for managing page and screen space
Commonly used layouts in multimedia page design
Design tips when using page templates and style sheets
printed pages. In multimedia, the concept of page layout has migrated metaphori-
cally from the kinds of physical pages that you can touch and turn, to digital
pages that appear on a computer screen or monitor. The page metaphor was
extended into the field of web design, but can also be more broadly applied to
any type of screen space where visual information must be arranged and ordered
for presenta- tion. In multimedia work, the principles of page design can be
applied to the cre- ation of a DVD menu screen, a full-screen title graphic in a
television commercial, or to the welcome screen of a micro app running on a
tablet computer or smart phone. Anytime you combine visual content within
the fixed or fluid space of a digital page or screen, you are engaging the activity
of page design (see Figure 5.1).
FIGURE 5.1
The DVD menu system
for the Walt Disney
film The Rookie
includes a main
menu, a scene
selection submenu,
and a bonus materials
submenu (pictured
here). The visual
layout of each menu
is the product
of multimedia page
design, where diverse
visual elements such
as video, graphics,
text, and animation
are arranged within
the screen space of
a 16 9
television frame.
While not
a visual element,
audio is also present,
adding to the
richness of the
Multimedia Page CHAPTER 5 129
multimedia Design
experience.
130 SECTION 2 Multimedia Design
densely populated with text. While early print- semper. Quisque ut lectus ac magna congue dignissim ac in nisl. Duis rhoncus elit at est tempor eu gravida est placerat. Vivamus aliqua
S t ro n g F o l lo w
lorem molestie.e id dolor arcu. Lorem ip um dolor si t a m e t, co nse ct et u r ad ipiscing
ing systems, like the Gutenberg press, could eas- FusceCurabitur
elit. Quisque at coat odio. id dolor aliq
arcu.etLorem
consequat odio. Curabitur
sum dolor
vulputate magna sit rutrum.
sit amet amet, consectetur
Nullam
uet vulputate magna sit amet
ad
ily mass-produce the transfer of letterforms on nec quam et metus b andit dictum ut id mauri s
.
Area
paper, they were not equipped to reproduce art- bh, semper
Donec sem ni h, semper tempus
ts rhoncus sit rhoncus
tibulum sem nec cursus.
sit auctor
met, aliquam met, aliquam
massa. Proinauctor massa.
consequat
a ursus dictum volutpat. Duis
ves ibulum sem nesus. Cras ursus dictum volutpat.
flow of printed text. If artwork was included, it was Nam lacinia nunc utnc ut suscipit
sem sem suscipit
ullamcorper. Vus tincidunt lorem et nisi co sequat dapibus. Aliqu m ut aliquet odio. Sed ut ipsu sita
drawn by hand after the pages were printed. am ut aliquet odio. Sed ut
eget ipsum tincidunt ut
r ivamus tincidunt lorem et nisi
m sit amet ante euismod tempus.
at leo. Nulla ligula dui, lacinia t orci r Vivamus at
F-Layout
The F-layout is one of the most common page layouts on the Web. As the name
suggests, the reader’s gaze is directed though the page in a pattern that resembles
the letter F (see Figure 5.3). As with most layouts, this one encourages the
user to begin reading in the top left part of the page. He then proceeds to
make a full scan from left to right across the first row of visual information.
Multimedia Page CHAPTER 5 131
Next, the Design
132 SECTION 2 Multimedia Design
user makes a partial scan of the second row. This pass rarely extends to the
end of the line or column. Each subsequent row of information is scanned in
short bursts from left to right as the user settles into a normal rhythm and
pattern for processing the main content on the page.
FIGURE 5.3
The Library of
Congress homepage
features the classic
F-layout design. A
prominent vertical side
bar along the left edge
of the page serves as
a visual anchor for the
eyes as the user scans
from left to right in
progressively shorter
bursts from top to
bottom.
Multimedia Page CHAPTER 5 133
Design
Humans are creatures of habit. For this reason, the F-pattern is an all-too-familiar
format for people who spend any amount of time in front of a computer surf-
ing the Web. We have grown accustomed to the most important information
being placed along the top edge of the page, usually in the form of an eye-
catching masthead or banner graphic. The masthead provides important
visual cues about the contents of the page. It can be scanned quickly and will
often influence a user’s next step. Will he continue scanning or bail from the
page? The second row of a web page often contains navigational prompts
like menu buttons or hyperlinks. If the first scan did its job by hooking the
user’s atten- tion, then the second scan will take on more meaning as he
seeks to understand how the page and site contents are organized. A sidebar
along the left edge of a page often comes into play on the third scan. It pulls
the user’s gaze towards the left edge of the document window, directing him
downward into a series of shorter bursts of scanning activity. Research shows
that conforming layouts to the F-pattern will enhance the usability of the page,
making it easier for users to glean information in an expedient manner.
FIGURE 5.4
The Smithsonian
Z-Layout homepage pictured
A less popular format, the Z-layout is a variation of the Gutenberg Diagram here guides the
(see Figure 5.4). Scanning begins with a full visual sweep across the top row of user’s eyes through
the page along a Z-
the page. The second scan flows diagonally through the center of the page in shaped path. Can you
locate other examples
of the F-layout and Z-
layout in action?
134 SECTION 2 Multimedia Design
a downward movement towards the bottom left-hand corner. The final scan is
a horizontal sweep across the bottom of the page. Important visual elements
should be placed along the path of the Z. While the Z-layout suggests a rigid
zigzag path, the chiseled angles of the Z-pattern can be softened to produce a
more organic S-curve feel. The Z-layout works best when a large visual element
or content region (like a single photo or visual montage) is placed in the cen-
ter of the page, between two rows of linear information. It is a simple layout
to create and can be used effectively for pages containing only a handful of key
elements that need to be communicated quickly without a lot of visual
clutter.
GREAT IDEAS
Visual Hierarchy
The term visual hierarchy refers to the perceived ordering of content within a page by the reader. A pag
Image
Heading 2
Image
Heading 3
Headings
The next thing you can do to bring order and structure to the presentation of text
is to add section headings and subheadings. A heading is a short descriptive title
or subtitle used to mark the beginning of a paragraph or content area. Maga-
zines and newspapers use headings routinely to set stories apart and to
provide readers with a brief informative prompt as they scan the page for
what interests them most. In a newspaper, the most important story of the day
is given visual prominence by placing it on the front page, above the fold,
and under a special type of heading called the headline. The headline carries
the biggest and bold- est font style on the page. Less significant stories are
attached to progressively smaller headings or subheadings.
In web page design, HTML specifies six levels of headings, identified with
the tags <h1>, <h2>, <h3>, <h4>, <h5>, and <h6> (see Figure 5.6). Head-
ing 1 is the largest and is often used for the masthead (master heading) to
identify the title of a page. Sectional content areas often begin with heading
2 or heading 3. Subsections within each area can be further delineated with
headings 4 to 6. To distinguish them from body copy, headings are set in a
larger font and weight, typically from 24pt down to 10pt in steadily declining
increments. However, headings can be customized to any size. They can also
be assigned a custom typeface, color, or border (see Figure 5.7). It’s important
to keep heading styles consistent within the page and across a series of con-
nected pages. For example, heading 1 should look the same on page 1 as it
does when reused on pages 2 and 3; doing so reinforces the design principle
of repetition.
136 SECTION 2 Multimedia Design
FIGURE 5.7
Headings are often
custom designed
as a part of the
overall theme of
a page layout.
The principles of
alignment, proximity,
and repetition are
critical ingredients for
creating a meaningful
and aesthetically
pleasing visual
hierarchy.
Multimedia Page CHAPTER 5 137
Design
GREAT IDEAS
The Golden Ratio
The Golden Ratio is a mathematical construct that has been widely used for
determining the proportions of a visual space. It can be seen in architecture, painting,
printing, and photography, and more recently in new media and the Web. Also
known as the Divine Proportion, the ratio is defined by the numerical constant Phi or
1.61803. As a ratio, it suggests that the optimum visual space is a rectangle whose
height is slightly longer than one and a half times its width (see Figure 5.8). The
vast majority of books printed prior to 1800 conformed to the Golden Ratio, which
served for centuries as a near universal standard for page proportions. Whether
horizontal or vertical in orientation, humans tend to favor visual pages and screens
that closely
mirror the Golden Ratio. For this reason, it is still embraced by many today, albeit more
loosely. The Golden Ratio has greatly influenced the design of page and screen
spaces used today in multimedia design.
FIGURE 5.8
The spiral form of a nautilus shell closely conforms to Phi, the golden number.
FIGURE 5.9
Staring at a blank page is a bit like walking into a house without walls. Where to start? For a designer, the first step is
to subdivide the vast unitary space of a blank page into smaller editable regions that can be individually arranged and
populated with content.
FIGURE 5.10
The design of a new web page often begins with simple sketches like this one showing how the
screen space will be broken apart into various content regions.
FIGURE 5.11
Graph paper can be used to break up a page into a variety of subspaces.
Multimedia Page CHAPTER 5 141
Design
Tech Talk
1
Samara, T. (2002). Making and Breaking the Grid. Beverly, MA: Rockport Publishers.
FIGURE 5.13
By altering the
grid structure, a
design space can be
subdivided in an
infinite number of
ways.
the page. Thus it is very important to have a vision and plan on the front end
for how a page will look when completed. The size of text and graphics needs
to be established early on to ensure that the columns and spatial zones
formed by the grid are made sufficiently large or small enough to hold the
intended contents. Proper planning is essential! A prudent designer will
rarely embark on creating a grid without giving long and careful forethought
to the intended structure and contents of the page. With this word of caution
established, let’s take a look at three of the most common grid structures used
in page design: (1) the single-col- umn grid, (2) the multicolumn grid, and (3)
the modular grid (see Figure 5.13).
SINGLE-COLUMN GRID
The simplest grid system features a single-column of visual information bor-
dered by margins on either side of the page. Also called the manuscript grid, this
layout is commonly used for document processing and book publishing,
where the content consists mostly of continuous text and few images. It is the
simplest grid to design. In fact, many programs, such as word processors, use
the single- column grid as a default starting point when a new document is
created. Because single-column formats are most often associated with high-
density text-based layouts, they are also most susceptible to the effects of the
Multimedia Page CHAPTER 5 143
Gutenberg Diagram. Design
144 SECTION 2 Multimedia Design
MULTICOLUMN GRID
A multicolumn grid breaks up the page with vertical divisions from left to
right, providing greater flexibility for integrating text and graphics within the
page. Multicolumn grids come in many varieties, with two- and three-column
formats being the most popular. Establishing a visual hierarchy is made easier
in a multicolumn grid because columns can be assigned different roles for
handling different kinds of content. For example, on a web page, a narrow
sidebar column might be used for small snippets of visual information such
as links, menu buttons, and short chunks of text that can be quickly scanned
by the user. A wider column might contain running text for displaying the
main content of a page. Columns can be equal in width or different. They
can be independent of other columns on the page, or connected, allowing
text and images to spill over from the end of one column into the next.
With a multicolumn grid, text no longer needs to flow in an uninterrupted
stream of continuous lines from the top of the page to the bottom. Columns
can contain negative space, headings, and illustrations, allowing text to stop
and resume less predictably at various points in the page. All things being
equal, reading text on a computer monitor is improved when it is format-
ted into columns, as opposed to running across the full width of the screen.
Doing so changes the start and stop points in a line, reducing the distance that
the eye has to travel back and forth with each scan. For smaller screens, like
those on smart phones and book readers, single-column layouts tend to be
more common.
Tech Talk
FIGURE 5.14
An inline photograph pushes the text downward (left), resulting in an undesirable page layout with too much white space. A more pleasing
MODULAR GRID
Modular grids include consistently spaced vertical divisions from left to right as
well as horizontal divisions from top to bottom (see Figure 5.15). With this grid
structure, designers are free to create spatial zones, modular design spaces
that can span across the width of multiple columns. Spatial zones add
horizontal flowlines to a page by subdividing column spaces into rows. With
spatial zones, designers can break free from a purely linear flow of content in
either direction. Modular grids are the most complex and difficult to
construct. However, they also tend to be the most visually dynamic and
interesting to work with.
Tables
A table is a rectangular grid consisting of editable regions called cells. Similar to
a typographic grid system, cells are formed by the intersection of columns
and rows. However, this is where the similarity ends. A typographic grid is
purely a conceptual framework underlying the structure of a page. It is a fixed
foundation
146 SECTION 2 Multimedia Design
FIGURE 5.15
Observe how headings
and a modular grid
work together to break
up the presentation
of visual information
in this mockup of a
traditional newspaper
layout.
Multimedia Page CHAPTER 5 147
Design
of modularity that will not change as the page structure on top of it is formed
and modified. A table, on the other hand, is a physical part of the page. It is
an algorithmic component created and modified from within the page-
authoring program. A table is a fluid structure that expands and contracts to
accommodate the placement of content within its cells.
Tables can be subdivided into row groups and column groups. Prior to the
development of Cascading Style Sheets (CSS), a topic discussed in Chapter 7
“Web Design,” tables were routinely used for structurally dividing the area of a
web page. While the World Wide Web Consortium (W3C) no longer encourages
this practice, tables can still be used “to arrange data—text, preformatted
text, images, links, forms, form fields, other tables, etc.—into rows and
columns of cells” within a page or regional subsection.2
DEFINING TABLES
A table is defined by how many columns, rows, and cells it contains. A table
with one column and one row (11) produces a single cell or region for storing
visual information. A table with three columns and four rows (34)
produces twelve cells. The more columns and rows in the table matrix, the more
cells you have to work with.
Table 5.1 A Simple 23 Table With Two Columns and Three Rows
Cell #1 Cell #2
Cell #3 Cell #4
Cell #5 Cell #6
TR/html4/struct/tables.html.
148 SECTION 2 Multimedia Design
1. Align left
2. Align Center
3. Align Right
4. The height of this cell
automatically increases as more
content is placed into the cell. 5 6
Lorem ipsum dolor sit amet,
consectetur adipiscing elit. Aenean
eu tortor sit amet libero placerat
molestie at sit amet metus. Ut
malesuada dolor sed leo imperdiet
feugiat.
7
8 9 10
As you can see, the contents of each cell can be independently formatted.
Cells 1–4 feature standard issue black text on a white background. In cell 5,
white text is inverted on a black background. Text alignment, styles, padding,
and borders can be customized at the tabular or cellular level, creating many
possibilities for visual organization. By default, table borders are turned on.
Keeping table borders visible in a page layout can draw unwanted attention
to the underly- ing structure of the grid. Turning off borders masks the grid
and often helps to reduce distractions caused by overt visible divisions. Table
5.3 shows a simple three-column layout using a 32 table structure.
Multimedia Page CHAPTER 5 149
Design
Tech Talk
The CSS Box Model The CSS Box Model was created to
border is typically a solid line, but other line styles can
give web designers greater control over the
be specified (dotted, dashed, groove, etc.). The margin
presentation of visual elements in a page. In CSS,
edge surrounds the box, and its width determines how
whenever a text or image element is added to the page,
much white space is placed between the outside of the
it is placed with in a definable spatial area or box. The
box and adjacent elements within the page or spatial
box has property values that can be adjusted to alter the
zone. A CSS box value can be applied uniformly to all
padding, borders, and mar- gins around the element.
four sides of an element or individually to one or more
Padding refers to the amount of white space between
sides at a time. For example, to insert a horizontal line
the content’s outer edge and its bor- der. With padding
beneath a box element, such as a heading or caption, the
set to 0, the content edge will be aligned automatically
border property can be set to “medium” for the bottom
with the border. The border property can be set to thin,
edge of the box and to 0 for the left, right, and top edges.
medium, or thick, or to any specific width. The
FIGURE 5.16
The CSS Box Model.
PAGE TEMPLATES
Unless you are an experienced graphic designer or page compositor, you
prob- ably lack the skills needed to design complex page layouts from
scratch. For- tunately, many programs come with layout templates or style
sheets that can be used for forming the grid system of a page. A page
template provides the structural divisions, color scheme, and general
formatting for a page layout and its contents (see Figures 5.17 and 5.18). The
content regions of the template are initially filled with placeholder text and
images. Placeholder material will eventually be swapped out with the
template user’s actual content. Nonsensical Latin sentences and paragraphs
are often used for placeholder text (also called filler or dummy text).
With templates, you often get what you pay for. The templates included with
your software, or those downloaded for free from the Internet, are often not
as well designed or sophisticated as the ones you can get from commercial
design
150 SECTION 2 Multimedia Design
companies. While professionals tend to shy away from using templates in com-
mercial work, sometimes there just isn’t enough time or client money to provide
a custom design from start to finish. If you have to use a template, spend some
time researching your options. Commercial templates are not too terribly expen-
sive and can offer you many more choices and options. Also, resist the tempta-
tion to blindly accept the color scheme and formatting included in the template
by default. Most templates give you some leeway in customizing the theme and
contents of the page. Take advantage of this by pushing the template as far as
you can. Remember: if you found the template, then someone else likely found it
as well. To make your version of the template stand out against the work of
oth- ers, you will need to take time to reshape it and rework it to fit your own
design preferences.
FIGURE 5.17
A web page template included with Adobe Fireworks CS5.
152 SECTION 2 Multimedia Design
FIGURE 5.18
The template shown in Figure 5.17 is transformed by replacing the placeholder contents (images and text) with original
assets designed for the project.
Multimedia Page CHAPTER 5 153
Design
FIGURE 5.19
WordPress.com is a popular blogging tool and web-based content management system (CMS). Users can choose a page
layout design from over 150 templates in the themes library.
People return often to their favorite websites expecting the content holdings of
the pages to be refreshed since their last viewing. Facebook is a good example.
As a social media site, Facebook uses CMS technologies to deliver dynamic
page content to viewers around the world. The vast majority of the dynamic
content (posts, photos, videos, etc.) is user-generated. However, the
advertisements are also dynamic, and typically, customized to the gender, age,
and profile of the user.
FIXED LAYOUTS
In the printing world, pages have a predetermined fixed width and height.
For example, a printed book, newspaper, or magazine has physical
dimensions that can be described by absolute units of measurement (inches,
centimeters, picas, etc.). Multimedia pages can be fixed as well, but given
the proliferation of so
154 SECTION 2 Multimedia Design
FLUID LAYOUTS
When a web page is defined by pixels, the size of the page will vary with the
resolution of the client monitor used to view the page. If the client is using a
small monitor, or if the display unit is set to a particularly low resolution
such as 800600, then any page layout that exceeds this size (e.g., 12801024)
will spill off the screen, forcing the client to scroll horizontally. A fluid (or
liquid) layout fixes this potential dilemma by using percentages instead of
pixels to define page size. When a fluid layout is used, the client’s browser
will scale the page to fit within the document window at whatever percentage
is specified. It is resolution-independent. For example, a page that’s set to a
width of 100% will fit edge-to-edge within the client’s browser window.
With the width set to 80%, the page will shrink, leaving 20% of padding
around the outer edge. Fluid layouts can be tricky to work with, but if used
correctly, they can lead to a more consistent viewing experience for the client.
With fluid layouts, the main content of the page will appear “above the fold”
as intended and horizontal scroll bars will be suppressed.
3
W3Schools. Higher Screen Resolutions. (2011). Retrieved from http://www.w3schools.com/browsers/
browsers_resolution_higher.asp.
Multimedia Page CHAPTER 5 155
Design
CHAPTER SUMMARY
Page design is an important part of the multimedia design process. While web pages
tend to garner the lion’s share of attention in the field of multimedia design, the
general concepts of page design presented in this chapter extend much further.
They apply to any activity related to the arrangement of visual elements within the
shared space of a multimedia page or digital screen.
156 SECTION 2 Multimedia Design
CHAPTER 6
Interface Design and Usability
by Qian Xu, Elon University
153
A picture is worth a thousand words. An interface is worth a thousand Key Terms Above the Fold Acces
pictures. Collaborative Filtering Cosmetic C
Data Mining Dropdown List Dropdo
—Ben Shneiderman, Computer Scientist and Expert on Forms Functional Customization H
Human-Computer Interaction (2003) Media Stickiness Menu
Motion Tracking Interface Navigat
Design is not just what it looks like and feels like. Design is how Tag Cloud Tailoring Text Field Thu
it works. Touch User Interface
—Steve Jobs, Co-Founder and CEO of Apple Inc. (2003)
Chapter Highlights
This chapter examines:
Types of user interfaces
The concept of user-centered design
The interface design process, from building a team to completing the interface
Commonly used navigational aids: menus, tabs, tools for managing hierarchy, and tools for organizing content
Techniques for designing better forms
Tailoring interfaces to users’ needs and wants through customization and personalization
Five components of usability: learnability, efficiency, memorability, errors, and satisfaction
Ways to improve the usability and accessibility of interfaces, including usability testing
USER INTERFACES
In the field of information technology, a user interface is any system that
supports human-machine interaction (HMI) or human-computer interaction
(HCI). This chapter adopts this broad definition of user interfaces to include
almost all information devices that a human being can interact with. The
interface has both hardware and software components and exists in the forms
of both input and output. The input component allows users to control the
system, whereas the output component enables the system to show the results
of user control.
Interactivity
Usability Usability Testing User Experience (UX)
User Interface (UI) User-Centered Design (UCD) Web User Interface (WUI)
Before we start to talk about different kinds of user interfaces, let’s clarify
one concept that is critical to this discussion: interactivity. Interactivity is one
of the defining characteristics of multimedia interfaces. In this book, we define
interac- tivity as the reciprocal communication between the media and the
users facili- tated by various technology features; in other words, it’s the user
and the media communicating with one another. As technologies advance,
user interfaces are becoming more and more interactive.
Interactivity implies both interaction and activity. It offers users—including you
—a new way to access and experience media content. It replaces the one-way
flow of information, such as with old-style broadcast television, with two-way
interaction between the media and its users. Instead of sitting on the couch and
passively viewing a television program, we can now consume media in a more
active manner.
By incorporating interactivity, we give users a chance to choose what they
need and want. They have the power to decide when, how much, and in which
format the information will be displayed. The ability to control media content
essen- tially makes users the source of information (see Figure 6.1). A user
has more agency when he or she can contribute to creating media content.
This increased sense of agency can boost “media stickiness” and better
engage users with the interface.
FIGURE 6.1
YouTube, the world’s largest distributer of video content, appeals to a new generation of
“interactive and active” users who want to produce, as well as consume media fare.
Source: Annette Shaff/Shutterstock.com
Interface Design and Usability CHAPTER 6 155
TOUCH INTERFACES
Touch interfaces use touch screens as combined input and output devices. They
let users directly interact with what is displayed on screen by using one or
more fin- gers. Touch interfaces can also sense passive objects other than
fingers, such as a stylus. The second screen on the lower panel of Nintendo DS
is a single-touch inter- face (see Figure 6.2). You can control the game by directly
using a finger or a stylus. The check-in kiosks at the airport and the check-out
kiosks in the grocery stores are also examples of single-touch interfaces. Apple
devices such as iPad, iPhone, and iPod further enhance the intuitiveness and
efficiency of touch interfaces by sup- porting multitouch, the ability to
simultaneously detect and accept two or more touch points (see Figure 6.3).
User-Centered Design
The goal of user interface design is to make user interaction as easy as possible,
so users can accomplish their goals in the most efficient way. This user-focused
approach is also called user-centered design.
User-centered design (UCD) is more of a philosophy and a process of
developing interfaces rather than a step-by-step method. It is about creating an
interface to meet the needs of real users rather than satisfying designers’ whims.
User-centered design does not require users to adapt to the new interface.
156 SECTION 2 Multimedia Design
Instead, it supports users by applying their existing behaviors and ways of
thinking to the new inter- face, so that they will have an intuitive and natural
interaction experience.
Interface Design and Usability CHAPTER 6 157
FIGURE 6.2
The Nintendo 3DS,
released in 2011,
features a single-touch
interface that users
can control using a
finger or stylus.
Photographer: Barone
Firenze/Shutterstock.com
FIGURE 6.3
The Apple iPad
features a multitouch
interface that
can respond to
multiple touch
points and gestures
simultaneously.
158 SECTION 2 Multimedia Design
Tech Talk
FIGURE 6.4
The Wii wireless controller (top) senses movement in three dimensions and provides haptic
feedback to users as they interact with the game. Below, an attendee uses body movements to
interact with a Kinect gaming system at an industry trade show.
Photographer: pcruciatti/Shutterstock.com
(Continued)
Interface Design and Usability CHAPTER 6 159
FIGURE 6.5
A teenager plays a video game with a wireless wheel remote.
Navigation
Good navigation is essential; its importance can’t be overstated. In
multimedia design, we often create complex products, more than we could
put on a single Web page or screen at one time. Navigation provides structure
and organization. It helps manage information. More importantly, it guides users
through the interface so that they know what information is available, where
they are, where they have been, and where they can go next. The most user-
friendly navigation is the one that doesn’t surprise users and allows them to
apply what they’ve already learned.
Based on functional differences, Anders Toxboe, a Danish Web developer,
grouped navigational features into several categories. In user interfaces, we most
commonly find those features that: (1) classify categories and create sections,
such as menus and tabs; (2) manage the hierarchy, helping users move within the
structure of an interface, such as home links, breadcrumbs, and footers; and (3)
organize content, such as thumbnails, carousels, pagination, archives, and tag
clouds.
MENUS
Menus are the most common navigational tool used in graphical user interfaces.
Space on an interface is limited, too limited to present all of your content. Menus
can solve the problem, letting you create links to sections with different kinds of
content. Unlike links buried in text, menus present links in consistent, easy-to-
find locations. If you are building a website or application with only a handful of
sections and little functionality, you might create a simple menu with a single
row or column of buttons or links. But more interactive menus—vertical
dropdowns, horizontal dropdowns, and accordions—can solve more complex
organizational problems and give your users a more interactive experience. For
Web pages, you can create these types of menus using HTML and CSS alone, or
you can create them using features in your Web design software, such as the Spry
Framework in Dreamweaver. Whatever technique you use, be sure to check your
menus and other features for accessibility (see the section “Making Interfaces
Accessible” at the end of this chapter).
FIGURE 6.6
Like many applications, the Safari Web browser provides a system of vertical dropdown menus for user navigation.
A keyboard shortcut hint is placed next to each command in the menu as a visible reminder to users. Amazon.com
features a horizontal dropdown menu system that pops out to the right of the sidebar when a user clicks on it.
Accordions
Accordions are a special form of menu with a vertically or horizontally
stacked list of sections (see Figure 6.7). Each section has a headline and a panel
associated with it (sometimes with links, sometimes with other information).
When a user clicks a headline, the section expands to show a panel with related,
more detailed information. At the same time, the other sections collapse,
hiding the other pan- els. By showing all section headlines but revealing only
one panel at a time, accor- dions can give users easy access to other sections
without causing information overload. If you use a vertical accordion, limit
the number of sections so users don’t have to scroll up and down to see the
complete list of sections.
Tabs
Designers use tabs when an interface needs to provide access to different content
modules or sections. This technique is similar to labeling folders and organizing
them in a file cabinet. Tabs provide an easy way to show a large amount of infor-
mation by category. When a user switches between tabs on the same Web
page, the Web page usually will not refresh.
Tabs are appropriate when users don’t have to see information in different tabs at
the same time: switching back and forth between tabs can create extra
164 SECTION 2 Multimedia Design
cognitive
Interface Design and Usability CHAPTER 6 165
FIGURE 6.7
This horizontal accordion component expands to reveal the contents of the section whenever
the user clicks on one of the six menu options.
FIGURE 6.8
The TrackMyPeople.com tabs are well-designed and follow all of the rules we’ve mentioned here.
Source: www.trackmypeople.com
166 SECTION 2 Multimedia Design
Home Links
No matter where users are, they always need to have the option to easily go back
home, back to the starting location in the interface. It might be a link
embedded in a website logo that brings users back to the homepage, or a main
menu option in the video game, or a home button on a mobile device
bringing users to the starting screen. Whatever it is, always place the home
link in the same location across different sections of an interface to let users
find it easily.
Breadcrumbs
Breadcrumbs are another such feature that can prevent users from getting lost
in an interface. The term breadcrumb comes from the Grimms’ fairytale Hansel
and Gretel. In the tale, two little children dropped breadcrumbs to form a trail
to find their way back home. Putting breadcrumbs in your interface gives users a
way to track their paths and go back up to another level in the hierarchy (see
Figure 6.9).
Breadcrumbs are usually used for complicated interfaces with hierarchically
arranged layers. If an interface only has a single layer, you don’t need bread-
crumbs. Breadcrumbs on websites are typically right below the navigation menus
in a horizontal line. They are often styled so they are plain and don’t take up too
much space. As a result, they have little impact on users’ processing of interface
content. Breadcrumbs can also be a good way to invite first-time users to fur-
ther explore the interface after viewing a landing page. Let’s imagine that a user
searches for an AMD motherboard and lands on a product list page on Newegg.
com. He or she fails to find a satisfying product on this page. However, enticed
by the breadcrumbs on the top of page, the user can navigate to a higher-level
product page to see what other brands Newegg offers.
FIGURE 6.9
Visitors to Newegg.com can quickly jump to a previous page by clicking on any of the links in the breadcrumb trail.
Source: www.newegg.com
Interface Design and Usability CHAPTER 6 167
Even though breadcrumbs may effectively reduce the number of clicks a user
needs to reach a higher structure on an interface, they should never replace
tabs or menus. They are not a form of primary navigation. They are just a
secondary navigation aid to help users establish where they are and to give
them an alter- native way to navigate the interface. Breadcrumbs should never
be placed on the highest layer of an interface, such as a homepage or an
introduction screen: after all, nothing is higher in the hierarchy.
Footers
You might have heard people say that all the important text and images should
be placed “above the fold,” so users don’t have to scroll down a page or a screen
to get to it. Extending this concept, the bottom of a page or a screen usually
becomes the part with the least important content. However, quite a lot of users
do scroll down to the bottom, depending on the site. It’s time to reexamine this
“forgotten” area and use it better.
In the footer of a website, you can add links to the frequently used sections
of the site, RSS feeds, social media sites, recent updates, or sign-up forms for
newsletters. Some interfaces even expand the size of the footer to add additional
internal and external links. For example, Last.fm has a fat footer that gives users
more choices of things to do after browsing the page (see Figure 6.10). It inspires
users to continue browsing. Rather than following the prescribed hierarchy of
the interface, the user is given the option to visit related pages using redun-
dant links embedded in the footer. The fat footer provides continuous access to
shortcuts to highly used sections, as well as major functions that are important
FIGURE 6.10
Last.fm incorporates a fat footer at the bottom of the page with additional internal and external links.
Source: www.last.tm
168 SECTION 2 Multimedia Design
GREAT IDEAS
Above the Fold Layout
Above the fold is a design principle that originated in the age of print media. Since most newspapers are
is relatively low (such as 640 × 480), the fold is relatively high. In contrast, when the resolution is rela
Organizing Content
The content of a user interface includes text, images, and video/audio downloads,
as well as other types of data. Let’s look at five popular methods of organizing
content.
THUMBNAILS
A thumbnail is a reduced-size version of a larger image (see Figure 6.11). It can
illustrate any visual item, such as a picture, a frame from a video clip, or a
screen- shot of a web page. If you have a lot of graphics for users to see at one
time, thumbnails should be your choice. They give users an overview of all the
visual content you want to present without requiring them to open all of the files.
This saves users download time and uses up less Internet bandwidth.
To make a thumbnail, you can create a mini version of the whole image. You can
also crop part of the image, but you have to be careful that users can recognize
the new partial image. In other words, it should also successfully convey the
mean- ing that you intend. For example, if you want to link to several great
French chefs with thumbnails, you would do better using their faces for the
thumbnails than their hands. Although some thumbnail sizes are fairly
common, you can adjust an image’s size to better fit your interface (see Table
6.1). You can even use an image that is partially transparent to give your
thumbnails a unique appearance.
CAROUSELS
Another way to organize visual content on a user interface is a carousel.
Carousels organize items on a horizontal line. Even though you might have
many items, the carousel shows only a few of them, as thumbnails, at a time. An
interface carousel is reminiscent of a carnival carousel with painted horses: as
the carousel turns, you get a good look at a few of the horses in front of you, but
you know there are more hidden to the left or right. If users want to see more,
they click an arrow (left or right), move the slide bar below the images, click to
Interface Design and Usability CHAPTER 6 169
the side of carousel indicators
170 SECTION 2 Multimedia Design
FIGURE 6.11
Clearleft uses unconventional stamp-shaped thumbnails to preview websites designed by them.
Source: clearleft.com
48 48 Very Small
64 64 Small
96 96 Medium
128 128 Large
144 144 Extra Large
160 160 Super Large
192 192 Huge
(the dots below some carousels), or mouse-over the images. To see an example,
look at the way that Time.com (see Figure 6.12 Top) introduces the topic of
addic- tion. Users can browse the various types of addiction by mousing over a
revolving carousel of small icons. As they mouse over the small images,
descriptive text with the statistics related to a certain type of addiction shows up.
Apple uses cover flow carousels, which have 3D effects, in iTunes and other
Apple products to present CD covers, photos, and other images (see Figure 6.12
Bottom and Figure 6.13).
Interface Design and Usability CHAPTER 6 171
FIGURE 6.12
A carousel interface is used in these examples by Time and Apple.
Source: www.time.com; www.apple.com
FIGURE 6.13
A carousel interface
option is built into
Apple OS X and can
be used for browsing
attached storage
drives and specific
file collections such
as your music library
from within iTunes.
172 SECTION 2 Multimedia Design
PAGINATION
You might use pagination if you have a large set of organized information, such
as a long article, a batch of customer reviews, a list of blog entries, or a group
of search results (see Figure 6.14). In these cases, you have to break the
information
FIGURE 6.14
Whenever possible,
add pagination to the
interface of multipage
or multiscreen sites.
Source (Top):
chriswalkerphoto.com
Interface Design and Usability CHAPTER 6 173
down to be displayed on separate pages. Unlike the simple “previous” and “next”
button, pagination gives users a brief idea about how much information to antici-
pate. It also allows users to easily navigate through the parts of a large body
of information. Users don’t have to finish browsing all information provided by
the interface. They can stop wherever they want and jump to other places. Do
you want to find out how that feature story ends? Just jump to the last page.
Good pagination is easy to use. First, you need to provide big enough
clickable areas to get to different pages and leave enough space between
page links. Sec- ond, you need to identify the current page—using color or
shape distinctions— and tell users where they are and where they can go.
Finally, your pagination should let users know the amount of available
information (number of pages). If you can, include both the first page link and
the last page link. However, if it’s hard to include both links, you need to add
an extra piece of information indicat- ing the total amount, as the second
example (bottom) in Figure 6.14 illustrates.
ARCHIVES
Another way to deal with a large amount of information is to organize it chron-
ologically using archives. Different from pagination, an archive is only appro-
priate for information that spans a rather long period of time, and that users
can logically browse chronologically, such as by date. Archive.org’s Wayback
Machine is an extreme example of archiving (see Figure 6.15). It does a great
job of capturing and keeping the homepages of all kinds of websites on a daily
basis. It organizes this information by year and month, using a blue circle to
mark days when it recaptured a site. Most sites don’t archive this much informa-
tion, so their designers don’t need to be as careful in planning their archive’s
time intervals. For example, if a blog has around 8 posts per year, it’s inefficient
to archive them by month. Another blog has 300 posts per year. However, these
posts may not be evenly distributed across months, with some months com-
pletely lacking posts. If this is the case, then the designer doesn’t even need to
list those empty months in the archive.
TAG CLOUDS
Tag clouds let users explore interfaces and organize content
unconventionally and give users agency (see Figure 6.16). They’re created
through folksonomy, or collaborative content tagging. You’ll see them on social
bookmarking websites and photo/video sharing sites, such as Delicious.com and
Flickr.com. Users tag pieces of content as they please, then the system
aggregates the tags and creates a tag cloud. As users add more tags, the cloud
automatically updates. Designers don’t control what tags users will apply to
content—hence, what words appear in the cloud—but they can make the
most popular or recent tags stand out by making them larger.
Tag clouds have an artistic feel. They are nice to look at but can be hard to use,
so don’t rely on them as your primary organizational system. Still, they give
users a different way to look at content, highlighting topics and issues
popular among other users.
174 SECTION 2 Multimedia Design
FIGURE 6.15
Which interface
would you rather
work with, the one
on top or the one on
bottom? Navigating
online databases and
archives can be an
arduous task. Making
the interface fun
and visual can make
the process more
enjoyable for users
and keep them
coming back to your
site.
Source: wayback
.archive.org
FORMS
You probably use electronic forms frequently: shopping online, logging into
a game system, registering software, downloading phone apps, and such. In
essence, you enter information and get feedback. Maybe your information
goes to another party—Amazon.com needs your order, payment
information, and
Interface Design and Usability CHAPTER 6 175
FIGURE 6.16
This tag cloud displays a collection of the “all-time” most popular tags generated by the users
of Flickr’s photo sharing website.
Source: www.flickr.com
address. But some forms just give you feedback, such as an inch-to-
centimeter conversion calculator. The bottom line to making your forms
successful is keep- ing them short and intuitive. The less you ask people to fill in,
and the easier you make the process, the more likely they will actually
complete and submit the form. Also, the more you avoid asking for sensitive
information or information they might not readily have on hand, the more
complete forms you’ll get back.
When you create a form, you can give users preset options to choose from with
checkboxes, radio buttons, and dropdown lists, or you can have them type con-
tent into text fields (see Figure 6.17). Here are a few options:
■ Checkboxes (small squares) let users select multiple options, even all of them,
if they’d like (e.g., interests: reading, playing music, playing sports, cooking).
■ Radio buttons (small circles) let users choose among mutually exclusive
options (e.g., yes or no; red, green, or orange).
■ Dropdown lists (text followed by a triangular down arrow) also let users
choose among mutually exclusive options, but you can reasonably include
more options in a dropdown list than with radio buttons (e.g., a list of
states to let you identify the state you live in).
Text fields let users type in what they want. Human beings are cognitive misers.
We try to fill out those online forms as quickly as possible. Sometimes, we might
not even bother to read instructions or labels when filling out a form. You can
help your users fill in the blanks with less effort in several ways: by using
autofill, prompt prefills, and tabbing. You should already be familiar with
autofill and autosuggestions: many email editors and search engines use them.
For instance,
176 SECTION 2 Multimedia Design
Last Name: *
Email Address: *
Home Phone: Cell Phone: *
Other Information
Gender: *
Female
Male Radio Box
Purchase Meals:
Monday Lunch Monday Dinner Tuesday Breakfast
Tuesday Lunch
Checkboxes
FIGURE 6.17
A typical form layout with text fields, dropdown lists, radio buttons, and checkboxes for user
input.
as you enter the first few letters of a contact’s name or an email address, some
email editors use autofill and automatically fill in a complete address. As you
type “multimedia” in Google.com’s search field, you get an autosuggested list
below the search box: “multimedia audio controller,” “multimedia fusion 2,”
“multimedia audio controller driver,” and more. And if you type two words,
“multimedia website,” you get more detailed suggestions of “multimedia web-
site template,” “multimedia website design,” and others. Autofill and autosug-
gestion work best for the most commonly searched terms. If you try to look for
something less common or rarely used, the system probably will not be able to
provide good suggestions.
How often do you read all of the text on a form you complete? Input prompts
prefill a text field with a prompt, telling you what to type. The website of Every-
Block New York City gives users a way to search for local news. The search
fields are prefilled with an example—“e.g., 175 5th Ave., Manhattan,
Williamsburg”— making the form field self-explanatory (see Figure 6.18).
Without such input hints, users might quickly scan a site and miss or ignore the
other hints below or next to text fields. But an input prompt that sits right where
you start typing will
Interface Design and Usability CHAPTER 6 177
FIGURE 6.18
Input prompts provide a visual hint to users entering data into a form field.
Source: nyc.everyblock.com
grab your attention right away. Ideally, you should also use a script that clears
the prompt text once your user starts to type. Also, if you design your forms so
users can tab through them, they won’t have to reposition the cursor to get to
each new field.
TAILORING
In the early days of computing, every visitor to a website received the same
inter- face. Web pages looked the same regardless who you were. Likewise,
applications displayed the same screen no matter who the users were. Some
sites and appli- cations are still designed that way, but many interfaces now
adapt to individual users, giving people tailored experiences. There are two
types of tailoring: person- alization, in which the system makes the changes, and
customization, in which the user makes the changes.
Personalization
When a computer system personalizes your interface, it’s using your previous
actions (sometimes information you put in a profile) to predict your interests.
In other words, it relies on your implicit interests rather than your explicitly
expressed ones. For instance, Netflix.com might recommend Avatar based on
the fact that you’ve rented other science fiction and fantasy movies. Facebook.
com examines your profile and gives you relevant advertisements. Sometimes
the process isn’t perfect.
A system needs data to make effective predictions and recommendations. It
might ask you to submit a questionnaire—collecting demographic, lifestyle, and
content preference information—when you register with a website or an
appli- cation. The system then applies rules and gives you a preset category
of content that fits your profile. Another way it can collect that data is by
data mining— such as with a website clickstream, or a record of your clicks—to
anticipate what action you might take next. Both these approaches require you
to log in to your own account. But it can collect data even if you don’t have
an account. Using cookies saved on your hard drive, the system can identify you
and your browser, then give you an individualized welcome message and
recommendations. But if you use a friend’s computer or a public computer
without first logging in to the site, you won’t get a truly personalized
interface.
178 SECTION 2 Multimedia Design
FIGURE 6.19
With every search, Amazon.com acquires valuable information about the shopping and buying habits of its customers.
Through collaborative filtering, this information is personalized and shared with other shoppers who are conducting a
similar search.
Source: www.amazon.com
Customization
Unlike personalization, customization allows users to deliberately tailor content,
giving them agency. As a user, you can directly choose options or even create
new content. As researcher Sampada Marathe explains, customization can be
either cosmetic or functional.1 Cosmetic customization lets you control
presentation, such as the background, color, font, layout, graphics, and your
profile photo. For instance, with a Wii, you can create an avatar of yourself,
selecting a gender, hair and skin color, facial features, name, and body type. At
the “feature” level of its interface, Facebook lets you hide your friend Bob’s
posts. And on iGoogle.com, you can change the background color and theme
of the site (see Figure 6.20). As the name implies, functional customization lets
you change the functionality of your interface. MyYahoo.com lets you add and
remove applications on your page and even drag-and-drop to rearrange the order
of applications. In Skype, you can create a specialized ringtone for contacts in
your address book. A ringtone might be a statement, such as “Mike is calling,”
a song, or some other sound.
Sometimes personalization and customization work together in one interface.
On YouTube.com, you can customize your own channel, changing its title,
creating tags, changing the background theme and color, adding and remov-
ing functional modules. At the same time, YouTube.com will also suggest vid-
eos that you might be interested in based on what you recently viewed.
1
Marathe, S. S. (2009). Investigating the psychology of task-based and presentation-based UI
customization. CHI Extended Abstracts, 3129–3132.
Interface Design and Usability CHAPTER 6 179
FIGURE 6.20
You can customize your iGoogle interface cosmetically by changing the theme and functionally by adding or
removing gadgets.
Source: www.google.com
USABILITY
Have you ever used a website that isn’t well organized or is complex but doesn’t
have a help menu? Perhaps you’ve used an online course management
system that requires extra steps to upload an assignment, a word processor
that loses a document because of one accidental click, a remote control with
more buttons than you actually need, or a touch screen that does not respond
quickly enough to the fine movement of your fingers. These are usability
problems. Designers can identify many of these problems—and fix them—
by focusing on usability during the design process: they apply usability
methods, including usability test- ing, to make sure their products meet
usability criteria.
Usability is a measure of people’s experience with a user interface. We can apply
the concept to anything a user uses, from software to hardware. Often, we apply
it to user interfaces because they are what users interact with. Usability refers to
the extent to which users can achieve their context-specific goals effectively,
efficiently, and with satisfaction. Can users successfully achieve their goals?
How much effort do they need to spend to complete a given task? And how
much do they enjoy the experience? In other words, usability is about the user. It
focuses on how to make the interface easier for people to use, rather than how to
train people to better adapt to a poorly designed interface. It’s hard to change the
habits of a large group of users, and if an interface is hard to use, many users will
find an alternative prod- uct or website that’s easier to use. In the end, it’s easier
—and more profitable—to make the interface more usable. Check for usability
180 SECTION 2 Multimedia Design
early and often!
Interface Design and Usability CHAPTER 6 181
Jakob Nielsen identifies five essential criteria that make something usable:
learnability, efficiency, memorability, error management, and satisfaction.2
Learnability
Learnability is how fast a user can accomplish the basic tasks on an interface that
they have never used before. An interface will be easier to learn if its design taps
into the core psychology of how we learn, and we learn well when something
is familiar, generalizable, predictable, consistent, and simple. If the design of an
interface is similar to the ones that users are familiar with, they will be able to
navigate through it by generalizing what they’ve learned in the past. They will be
able to predict what will happen when they click a certain button or figure out
how to go back to the homepage. If you keep the design pattern of the interface
consistent across different sections, users will be less confused and will need to
adapt their behavior less. Learning is fast when there is not much to learn. Less
complexity also can improve learnability.
Efficiency
Efficiency is how fast a user can perform tasks after learning how to use the
inter- face. While learnability is about users’ initial interactions with the
interface, effi- ciency is about the effort users expend during repeated visits.
For example, two shopping websites can be equally learnable if their check-
out buttons are well placed, making it easy for users to complete a purchase.
But one may be more efficient than the other. One may require a single step
for completing the pur- chase, while the other may take four steps and
require users to fill in several forms. An efficient website can ultimately save
users a lot of time, make them happy about their experience, and boost both
ratings and repeat business.
Memorability
If a user has used the interface before, can he or she still remember enough to use
it effectively after being away from it for a while? This is the issue of memorability.
It’s not surprising that users tend to forget how to use interfaces they only use
occasionally. If they use several different systems at the same time, such as three
different banking websites, they’ll have an even harder time. Relearning a
system every time they come back takes extra time and effort, and some people
may take their business elsewhere. Memorable feedback helps users, and
familiar visual cues make an interface both more learnable and memorable.
Error Management
Although usability is primarily focused on eliminating design mistakes rather
than blaming users, we realize that people will click buttons we didn’t expect
them to click and make other decisions we might think of as errors. A good
interface: (1) reduces the number of errors users might make; (2) decreases
the severity of each error, and (3) makes it easy for users to recover from
errors. If
2
Nielson, J. (2010). Usability 101: Introduction to usability. Retrieved December 2nd 2010 from
http://www.useit.com/alertbox/20030825.html.
182 SECTION 2 Multimedia Design
possible, avoid creating situations in which users are likely to make mistakes.
For example, if you expect users to look for content or a feature that’s on another
page, add a link to it on the current page. And if a user submits a form with
an empty required field, send the user back to the partially completed form—not
an empty one—and point out what’s missing.
Satisfaction
Satisfaction is about enjoyment. How pleasant is the user’s experience? How
much do users like the interface, and how satisfied are they with it? We can learn
about the other three elements of usability by watching users interact with the
interface. Not so much with satisfaction: to understand satisfaction, we need to
ask for users’ subjective feedback.
If you are designing something only you will use, you can do almost
anything. But most interface design is a creative activity bounded by
constraints: time, money, environment, and more. Usability is just another
constraint, an essential one. Checking usability lets you explicitly identify users’
needs. You learn how people usually perform on interfaces similar to what you
plan to design. It sheds light on real issues that you need to consider and
inspires ideas for improve- ment. Usability is never the enemy of creativity.
You, as the designer, ultimately choose the details of your design. But usability
principles and testing give you a way to ensure that people will actually want to
use your website, play your game, or visit your kiosk.
Usability Testing
If you want to ensure the usability of an interface, you need to conduct usability
tests early on and again when you develop a working prototype. They are the
best way to understand how users really experience your interface and where
your design isn’t intuitive. If you’re faithfully following the design process
we’ve discussed, you’ve already identified the target users of your interface and
devel- oped the prototypes mimicking the entire interface or the part of it you
are interested in. Now recruit your test participants: several users or good user
prox- ies. You can catch about 80% of the problems with your interface if you
choose five to eight good users and design your test carefully. You can conduct
usability tests in a lab or in a more natural setting, where the interface is
supposed to be used.
There are several ways to conduct a usability test. Let’s look at two: (1) unobtrusive
observation, and (2) think-aloud protocols, also called obtrusive observation.
Whichever method you choose, avoid bias by using a script (such as the Test
Facilitator’s Guide available on Usability.gov) so each participant receives the
same instructions when you explain tasks. Also, make participants comfortable
with the process. People often have anxiety about testing, even when there’s no
grade involved. Remember to explain that the interface is being tested, not them.
Tell them that any problems they have are problems with the interface and that
they are helping you when they find problems.
Interface Design and Usability CHAPTER 6 183
With each method, you usually give participants one or more tasks to
complete. With the unobtrusive observation method, after you give
participants the basic instructions, you observe what they do without talking
to them. You can record how much time they need to finish a task, how many
errors they encounter, whether they are confused by a certain function, as
well as their physical and emotional reactions. Even if you only have a paper
prototype, you can give them tasks to complete (e.g., find out how to email the
Vice President of Operations), have them point to where they would click, and
watch what they do. In contrast, when you use the think-aloud method, you
will not only keep a record of what they do, but also encourage them to talk
about their experience as they interact with the interface. Is it hard or easy?
What do they expect to happen when click- ing a thumbnail or image? If they
were to encounter a similar problem in a dif- ferent interface, how would
they solve it?
If you are testing a fuller prototype, give participants tasks that will require them
to use or interact with at least the following types of components:
■ main menu and submenus
■ interactive components (forms, search boxes, popups, and so on)
■ text and informative statements
■ images (especially icons)
Ask your users to go to specific pages or screens, use forms, complete tasks that
involve icons, and find answers to questions that require them to go to pages
or screens buried in your architecture. Ask them to do anything an average user
might do. If you have pages or screens that they can’t get to in one or two clicks,
choose some tasks that require them to go to these deeper spots. If you have
submenus, choose tasks that require the users to use them. For instance, if you
are designing a website for a club, and you have membership information that
they can get to only from a submenu, ask a question about membership (e.g.,
How much are dues?). Make sure your design isn’t missing features they want,
like search boxes, contact links, and such. Get their reaction to the colors, fonts,
layout, images, and other traits of your design.
With the think-aloud method, you can collect users’ responses and
understand their decision-making in real-time. You can also collect much of
the same data you would in unobtrusive observation; however, you might
interrupt the normal flow of users’ interaction with the interface, making the
interaction unnatural and some data meaningless. What will a users’ “time on
task” (the time it took them to complete a task) mean if you interrupted them
to talk about what they were doing at length?
In addition to observing users’ actions, you can carry out focus group
research, letting users sit together and discuss their experiences with the
interface. You can also collect their opinions in one-to-one interviews to get
more in-depth feed- back. Questionnaires, both pencil-and-paper and online,
are yet another way to collect users’ feedback after they finish interacting
with the interface. If you would like to compare two types of interfaces or
two functions of an interface, conduct an experiment.
184 SECTION 2 Multimedia Design
GREAT IDEAS
Accessibility and the Law
A number of government acts call for accessibility. We’ll talk about two. Section 508 of the Rehabilitation Act requires Fe
technology, such as websites, accessible to people with disabilities; this is frequently a requirement of organizations rec
GREAT IDEAS
Accessibility and Usability
Accessibility is related to usability. However, accessibility should not be confused with usability. Wher
preferences: reverse-contrast text and larger headings for low-vision users, color
schemes for color blind users, and more. And you can create separate style sheets
for mobile devices and printing.
The way that we interact with technology has evolved significantly through the years,
CHAPTER SUMMARY
and it continues to change. Motion tracking interfaces and multitouch interfaces prob-
ably will become more popular in the near future. Interface designers are also working
on developing more natural input devices, some controlled by a user’s natural eye gaze
or even by brainwaves. The display devices we use are becoming lighter and more flex-
ible, and upcoming devices will likely break some of our current conventions. Samsung
recently demonstrated a transparent display that could be used for MP3 players,
and LG has developed flexible e-paper that can be bent. In general, interfaces are
becoming more intuitive and giving us more interactive experiences.
No matter what type of interface you create, it’s essential to focus on real users
through- out the design process. When you plan navigation features, consider how
they will be used. When you design forms, make them easy to use. Consider tailoring
your interface to give users a better experience. Whatever your ultimate design, work
to ensure it is usable and accessible, not just trendy.
181
Key Terms Accessibility Absolute
The future of cyberspace is not really about technology. Although it will Attribute Audience
obviously continue to develop, the essential technology is already here Background Image Bandwidth Ben
in one form or another. The thing we have to start recognizing is that Breadcrumb Trail Broadband Brow
Button Button State
cyberspace must be content-driven rather than technology-driven. Cascading Style Sheets (CSS) Com
—Rick Schremp, Cyberspace 2020 (Govtech.com), (1995) Dial-up Domain Name
Drop-down Menu Element
File Transfer Protocol (FTP) Forms
Chapter Highlights Frames Head Headings HTTP
This chapter examines: Hyperlink Hypertext Markup Langu
The World Wide Web and Hypertext Markup Language (HTML) Inline Element Internet
The basics of HTML code and the technology behind client/server networks IP Address Local Area Network (LA
How to create and manage project files and website assets
The process for researching, planning, designing, producing, uploading, and testing a basic website
The importance of separating meaning from presentation and designing accessible sites for all audiences
Berners-Lee, T. (2000). Weaving the Web: The original and ultimate destiny of the World Wide
1
HTML
Web pages are built using Hypertext Markup Language. Rather than being a pro-
gramming language, HTML lets you put tags (markup) around your text so
that the browser knows how to display it on the screen, which words or other
objects are links, and where to go if you select the link, among other things. This
chapter talks about the latest version, HTML 5, which has new features such
as built-in support for playing media files, but mostly addresses those parts of
HTML 5 that
Web CHAPTER 7 183
Design
are compatible with HTML 4.01. That way, you can easily make your
websites available to people using browsers that don’t yet support HTML 5.
The newer HTML 5 will be a much more versatile standard. If you are
planning to work in the online industry, you will need to keep your
knowledge current, though the finalized version isn’t slated for release until
2014. In addition to adding new elements, HTML 5 should free developers
from having to rely so heavily on JavaScript and plug-ins such as Flash. Keep
tabs on the latest HTML 5 decisions on the W3C website; if the information on
that site is too technical for you, read the latest tech blogs and other well-
updated sites about HTML 5.
HTML files are just text. Although many people use web design software such
as Dreamweaver and Expression Web to create web pages, you can create them
with nothing more than a basic text editor such as Notepad in Windows. The
graphics used in a web page are not actually inside the text document, but are
referenced in the code and then uploaded along with the HTML file to the server.
Formatting the text works much like a word processor, assuming you had to type
in the commands to make things bold, italicized, etc., which once upon a time
you actually had to do. Most, but not all, elements have a start tag and an end
tag (also called opening and closing tags), which tell the browser to start doing
something and to stop doing something. Figure 7.1, shows one way to render
text in bold.
In addition to marking up how things look graphically, you can also mark up
the text semantically, giving different parts meaning, as in labeling some text as
a paragraph (<p></p>). Instead of marking something as being in bold
(<b></b>), you can mark it as being strong (<strong></strong>). The same is
true for italiciz- ing. You could mark up the text to be presented in italics
(<i></i>), but it’s better to mark it up semantically as emphasized text
(<em></em>). If you decide later to make text tagged with <em> bolded rather
than italicized, you can do that without changing the HTML tags.
Good designers try to use semantic—or meaningful—markup rather than
markup that controls presentation alone. So they opt for <em> tags rather
than tags like <i> and control how the emphasized text looks in a style sheet,
usually in a separate file. We’ll get more into that later. For now, picture labeling
your text in meaningful ways—with chunks marked up as headings,
paragraphs, lists, block quotes, images, links, and emphasized text—and
having a separate file control how it all looks. You might have a 500-page site,
but at a whim, you could change that one style sheet file and make all of the
paragraphs display in Arial rather than in Times New Roman without changing
anything in those 500 pages. Handy, no?
FIGURE 7.1
How tagging works.
184 SECTION 2 Multimedia Design
Tech Talk
FIGURE 7.2
HTML 5 sample.
This technique is a much better way to manage a big site. It also lets you
include a separate style sheet for mobile devices so users can easily get to
your informa- tion on their phones and tablets. Also important, it gives
visually impaired users who rely on assistive technology (including screen
readers that read pages aloud) an idea of what your text means. Instead of
marking up a piece of text so it’s presented as Arial 14 pt bold, you could
mark it up as a heading, and the screen reader would read it aloud as a
heading!
Look back at that sample of code above. Notice how the head and body tags
above are inside the HTML tags? This is called nesting. Have you ever had a set
of nesting dolls, the two-part hollow dolls that stack inside each other? Tags
work the same way. You have to make sure the start and end tags in one pair
of tags fit completely inside the start and end tags of another pair. Just like it
wouldn’t work if you tried to put the head of a small doll on once the bottom
is already sealed in a larger doll, as you can see in Figure 7.3, your HTML won’t
Web CHAPTER 7 185
Design
work if you don’t nest the tags properly. HTML 5 lets you leave the end tags
186 SECTION 2 Multimedia Design
FIGURE 7.3
Nesting HTML tags.
off of certain elements, but you can also include those end tags. Because older
versions of HTML require them for most elements—as do many elements in
HTML 5—consistently using end tags is a good practice, with exceptions.
Some types of elements often deal with bigger chunks of information than others
and usually begin on a new line. These are called block-level elements. Examples
include paragraphs, lists, and list items (the separate bulleted or numbered
items in a list). In contrast, inline elements such as strong tags, often control
small chunks of information and are used within a block-level element. In the
exam- ple shown in Figure 7.3, the paragraph element is block-level, and the
strong element is inline. You can nest some block-level elements in others,
and you can nest inline elements in block-level elements, but you can’t nest
block-level elements in inline elements (so you can’t put paragraph tags
between bold tags).
Let’s take another look at the parts of an HTML page. While the browser knows
it’s looking at HTML, without the doctype declaration, it doesn’t know which
type of HTML. The W3C periodically issues new HTML standards, and the
browser needs to know which version (doctype) is being used so that it can cor-
rectly interpret and render the page. If you don’t include a doctype, the browser
may not render the page correctly. Once all browsers support (correctly interpret
and render) HTML 5, you’ll be able to reliably use that simple piece of code:
<!DOCTYPE html>. Although people are beginning to use the draft version
of HTML 5, it is still being developed and refined, so this doctype isn’t
standard yet. HTML 5 offers new features, but the code in this chapter is
common to both HTML 5 and HTML 4.01, so you shouldn’t have any problems.
If browsers aren’t displaying your page correctly, try using the “strict”
doctype for HTML 4.01, a complicated piece of code that is best copied and
pasted right from a reliable source, such as W3C. Here’s what it looks like:
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.01//EN"
"http://www.w3.org/TR/html4/strict.dtd" >
Web CHAPTER 7 187
Design
Tech Talk
FIGURE 7.4
Anatomy of a basic tag.
FIGURE 7.5
Anatomy of a complex
tag.
Now let’s look at some body tags, which control what the user sees.
Although the tags in Figures 7.1 to 7.3 are simple, they get more complex.
There can be several parts to a tag, as with the hyperlink in Figure 7.4.
Some resources call the tag the element, but technically the element here is
every- thing from the start tag to the end tag and in between. HTML 5 will
streamline code, making it cleaner (less cluttered and more efficient) and
pushing devel- opers to control presentation in external style sheets. You will
find examples of older approaches to coding as you analyze pages to see how
other people made cool things work, but you are better off controlling styles
externally. In older code, especially, you might see examples of more complex
attributes as shown in Figure 7.5. Again, though, you are better off controlling
colors and other presentation characteristics from a separate style sheet file
rather than within the element.
Let’s look at a couple of things in the examples you just read. Did you notice that
every time you saw an equal sign, it was always followed by quotation
marks? This is a handy thing to remember as you start writing your own
code. You’ll probably read elsewhere that these quotation marks are optional
in HTML 5, but
188 SECTION 2 Multimedia Design
since you sometimes need them to make your code work right, get in the habit
of using them consistently. Also, did you notice that some of the code above was
indented? Indenting your code can make it much easier to read and brings up
another point: browsers ignore multiple spaces in HTML. If you put twenty
spaces between two words, it’s the same as putting one space between the words.
Brows- ers also ignore blank lines you put between your words, as shown in
Figure 7.6.
If the browser doesn’t see a command telling it to put extra space between the
words, it won’t put the space in. HTML is generally not case-sensitive; even
though the doctype is in uppercase, that’s only by convention. Among the excep-
tions are special character commands, which have to be lowercase. Using low-
ercase is the best practice because these exceptions exist, and because lowercase
is easier to read, most developers use lowercase, and the practice of using low-
ercase will prepare you to branch out into similar coding that is case-sensitive.
FIGURE 7.6
Blank spaces in HTML.
Tech Talk
FIGURE 7.7
Special characters in
HTML.
Web CHAPTER 7 189
Design
Browsers
One of the challenges to designing web pages is that what looks good in one
browser may not always look good in another—the browsers don’t always inter-
pret HTML and other elements the same way. And there a number of web
brows- ers: Firefox, Chrome, Internet Explorer, and Safari, to name a few. Not
only is code sometimes interpreted differently between browsers, but different
versions of the same browser may behave differently, and the same version of a
browser may interpret code a bit differently when run on different platforms
(Windows, versus Linux, versus Apple’s OS X). Part of the problem is that
HTML is an evolv- ing language. The meanings of HTML tags sometimes
change, new tags are added, and old tags are deprecated (made obsolete and
invalid). Watch out for browsers that don’t follow W3C standards. One of the
most notorious is IE 6, which often requires special coding. Although it came out
in 2001, there are still people using it, particularly outside of the U.S.
One thing most graphical browsers have in common is that they rely on plug-
ins to show non-HTML-based material, such as video, Flash, Adobe Acrobat,
and Microsoft Word files. Plug-ins allow you to run a program inside your web
browser and have become an integral part of the Web. When you are designing a
web page, you need to be careful about relying too heavily on content that
requires plug- ins. If your end user doesn’t have the plug-in, the page won’t work
correctly. Also, consider how many people browse the Web on mobile phones
and other devices, such as the Kindle. Many types of devices have browsers that
may not support all plug-ins. For instance, Flash isn’t currently supported by
Apple on the iPhone or iPad. As we move from HTML 4.01 to HTML 5, plug-
ins are becoming less neces- sary. That said, HTML 5 won’t be fully supported
for several years.
When you develop a page, you need to make sure that users of graphical brows-
ers will see what you want them to see, that mobile users can easily access
materials—and the page isn’t just a hard-to-read mini-version of the one you
designed for the person using a desktop computer—and that other users can too.
While most of us are familiar with graphical web browsers like the ones
listed above, some text-based browsers are still in use, including Lynx and
Bobcat. And you should also think about how users with assistive technology
“see” your page. We’ll talk more about that later.
The Network
THE URL: YOUR ADDRESS ON THE WEB
Communicating on the Web is a bit like communicating with postal mail—
assuming you had to mail off a request for information every time you
190 SECTION 2 Multimedia Design
wanted
Web CHAPTER 7 191
Design
FIGURE 7.8
Anatomy of a uniform
resource locator (URL).
INTRANETS
We tend to think about web content as being something we use on the
Internet; however, designers also build content for use on Intranets. Intranets
frequently work much like the Internet itself and rely on many of the same
protocols, including HTTP. Whereas the Internet is generally considered a
public arena, Intranets are private and are usually restricted to employees of a
company. In many cases, a company will restrict the use of its Intranet to
computers with IP addresses owned by the company. Users outside the
company’s local area
Web CHAPTER 7 193
Design
network (LAN) have to log in to a Virtual Private Network (VPN) server so that
they can get an authorized IP address and gain access to the company’s Intranet.
If you are in college, you may already be logging in to a VPN to get access to
resources not otherwise available from off campus (such as library
databases), or possibly to use the campus wireless network.
GREAT IDEAS
File Naming Conventions
How you name your files is critical. Here are some general guidelines:
The main page of your website should be called index.htm or index.html. Most web servers are set up to display a default file if
(Continued)
194 SECTION 2 Multimedia Design
Be consistent with which file extensions you use for your web pages. A file extension
is a suffix used to indicate a kind of file. Develop a religious-like devotion to .htm or
.html, the two file extensions that tell the computer you have an HTML file. You can
use either .htm or .html, but if you are not consistent, you risk ending up with both a
project.htm and a project.html file and changing or uploading the wrong one. You can
run into similar problems with naming JPEG graphic files. You’re better off using the
more standard .jpg instead of .jpeg.
Keep your file names lowercase. Lowercase file names tend to be easier to read. In
addition, most web servers can tell the difference between uppercase and lowercase
file names. Some servers are case-sensitive: even if you only have a Projects.htm file,
the server may not find it if the user types the URL in lowercase. Also, you don’t want
to get caught in the trap of having a projects.htm file and a Projects.htm file and linking
the wrong one.
Don’t put spaces in your file names. Most web servers will not display spaces in a
file name. Instead, they will fill the space with a %20, so the file name “my
projects.htm” ends up being “my%20projects.htm.” Not only does this look ugly, but it
also makes it difficult for users to type in a file name. If you feel compelled to put
spaces in, use an underscore: “my_projects.htm.”
Don’t use characters that aren’t a letter, number, underscore, or hyphen. Such
characters may cause problems with some servers and can make it difficult for people
to remember or type a file name. You should also avoid using non-English letters
such as ë.
Avoid extra periods. For example, my.project.htm. Periods in file names are generally
used only to separate the name from the file extension.
Many of the above guidelines apply to non-HTML files as well, including graphics and
video files. Some organizations have their own naming conventions. When you are
working for others, make sure you learn their naming conventions and use them.
Establishing a Hierarchy
Most websites have a hierarchy, often with three main page levels, as shown in
Figure 7.9:
■ home (your default page, index.htm or index.html)
■ main pages (linked from a main menu)
■ content pages (linked from local navigation and from other content pages)
A first step in site design is deciding how many pages you need to link to in
the main navigation on your home page and whether you’ll need a submenu
(also called local navigation). Your navigation area needs to be big enough to
accom- modate your major links. As you plan what users will click on to
reach their destinations, create a chart to figure out how pages can be
reached. Try to plan your site so users can reach any page with three or fewer
clicks from the homep- age. Main menus usually run along the top of the page
or the left-hand side of the page. Users are used to navigation being in these
areas. Avoid the right and bottom for main navigation.
Web CHAPTER 7 195
Design
Home
FIGURE 7.9
Youth Community Parents Article 1 Article 2 Article 3
A hierarchical
site map.
Main menu
Page title Content: Pictures
Footer
FIGURE 7.10
The box layout.
Web CHAPTER 7 197
Design
FIGURE 7.11
198 SECTION 2 Multimedia Design
Coding the box layout.
Web CHAPTER 7 199
Design
the box for your main text or other information is <div id=“section”>. You’ll
notice that all three of these are contained in a larger box: <div id=“container”>.
The properties for each div, such as its position, margin, background, or font
characteristics, will be controlled by setting the properties for that id in a style
sheet. The id attribute assumes that the label will only be use once on any given
page. If you need to reuse a label more than once on a page, you need to use
the class attribute. For example, if you were to add another div that you wanted
styled the same way as the one labeled “section,” you would need to use a class
for each, not an id: <div class=“section”>. We’ll talk more about class and id and
style sheets a bit later in the chapter.
The boxes in Figure 7.11 are all made with generic div tags, and all current
brows- ers will recognize them. Notice how, without the indenting, it might be
hard to see which div an end tag closes? It gets harder when you make more
complex pages and fill them with content. As HTML 5 becomes standard, move
toward using semantic tags for most of these boxes (e.g., <nav> and <header>,
<article> for syndicated- style articles, <section> for general content, etc.). For
now, use divs and label them semantically (e.g., <div id=“nav”>) so you can
transform them easily in the future, and so you can figure out which is which.
Learn how to add HTML comments, like the one in Figure 7.11, to label the div
end tags and add other important comments.
Notice, too, how the header, nav, and section divs are all nested inside the wrap-
per div? This structure gives you control, in CSS, to position the main part of the
site as a whole (all the items in “wrapper”), perhaps against a larger color
back- ground that fills the whole browser, and also to position items within
“wrapper” in relationship to the larger div and to each other. Without any
fancy position- ing, the items would show up in the normal flow from top to
bottom, just as they are placed in the code. With a little CSS coding, you could
move them around in relationship to one another—for instance, with the menu
on the left—without changing your HTML. That’s beyond the scope of this
chapter, though.
When you pull up a web page in your browser, you’ll see page title (<title>
</title>) information above your browser’s buttons. The page title, controlled
in the head, is especially useful if you have multiple pages open, because
each window or tab shows the page title, helping you find the right page
when you move back and forth. The title makes the page more usable.
Choose a mean- ingful title. Also, either choose a short title, or choose a title
that has the basic site information in the first word or two: when titles appear
in tabs, you’ll usu- ally see only the words at the beginning of the title. For
example, if you were shopping for shirts online and had five pages open, a
title that starts with the store name would be more useful than a title that
starts with the word “shirts.” Perhaps the title will include a pipe (|)
separating the company name from the page name.
and headings. You may cause other problems, too. Again, some elements are
block-level and others are inline.
Headings
You will probably use headings throughout your website, much like a
textbook uses headings. Headings give users a sense of the structure of a site at
first glance, and assistive technology allows users to review the headings on a
page without wading through all of the content. HTML gives you six options for
heading sizes: h1 (the largest because it is the first level heading) through h6
(the smallest). You can define how these headings look by using a style sheet,
but the headings will be easy to read and tell apart even if you don’t use a
style sheet. Headings automatically appear with space above and below
them.
Lists
Another useful way to mark text is with list tags. You may choose a bulleted
list (an unordered list, or <ul>) for items that don’t have to go in a particular
order, or a numbered list (an ordered list, or <ol>) for items that are sequential;
each uses list items (<li>). Alternatively, you might use a definition list (<dl>),
in which each item is broken into two pieces, the definition term <dt> and the
definition data <dd>. Figure 7.12, outlines the structure of each type list.
FIGURE 7.12
The three types
of lists.
202 SECTION 2 Multimedia Design
FIGURE 7.13
Linking content on the
page and linking files.
Web CHAPTER 7 203
Design
FIGURE 7.14
A sample file structure.
FIGURE 7.15
Navigating directory
paths in HTML.
Let’s say you have everything in a folder named My Work. When you
upload your site, you’ll upload the contents of My Work, but not the folder
itself. For now, you’re working on the site offline (not the version on the
server). A few files would be in that main folder: your homepage (index.htm)
and your main pages (about.htm, portfolio.htm, design.htm, and
contact.htm). Notice that the site has only one index.htm file, and it’s in the
main folder. That folder con- tains other folders with simple names to
organize deeper site content. If this site had images, it would probably have
an images folder. Figure 7.15 shows how to structure links between some the
files shown on the right of Figure 7.14,
204 SECTION 2 Multimedia Design
Tech Talk
Navigation bars can be simple or complicated. You can use other types of code,
like JavaScript, to create navigation bars, but some of these scripts slow down
the browser or pose accessibility problems.
FIGURE 7.16
Linking images and
using alternative text.
Web CHAPTER 7 207
Design
graphic button. Even if you create a site for visual artists and expect only sighted
users, some images may not be rendered. Do you recall pulling up a page and
seeing a box with an X or an odd error icon in it where the image should have
been? That’s a rendering problem. To make sure that content is still usable, add
an alt attribute to your img element to describe it.
Creating Tables
Originally, HTML tables were designed to hold tabular data—and that is a great
use for them. Some designers, unfortunately, use them to lay out web pages
by turning off the borders and using the cells to place images, colors, and text
right where they want. Don’t fall into this trap. Tables are not as flexible as divs:
they make your content harder to scale (harder to magnify and shrink) and
your job of preparing the page for a different presentation—such as on a
mobile phone—tricky, and they frustrate users who rely on assistive
technologies like screen readers. (Among other things, imagine having to
hear about the table as you listened to a web page being read: “Table with
two columns and four rows. Welcome to … ”). Tables are still used to hold
tabular data, though, so you might use one to show a menu of upcoming yoga
courses and their prices at a gym.
Tables are built row by row. Essentially, you tell the browser you want a
table; within that table you want a row; and inside that row you want a
cell, another cell, and yet another cell. Then you add another row and con-
tinue. Imagine listening to a screen reader read these cells aloud as they
are written in the code. You could create quite a mess for your users. So as
you develop a table, make sure you build it correctly and use semantic, or
meaningful, markup so assistive technology can help readers make sense of
the information. Good tables include a summary attribute in the <table>
tag, start with a caption, and are then built by row groups in up to three
sections in this order:
■ Table rows <tr>. In your table head, your row will contain the next elements,
the table headers. In your table foot and table body, your rows will
contain table data (normal cells).
208 SECTION 2 Multimedia Design
■ Table headers <th>. You can (and should) add a header to each column to
label what comes below.
■ Table data <td>. Your cells (other than headers) are in table data elements.
Let’s look at a simple table in Figure 7.17. You’ll notice that the width is set
to 400 pixels, and the border is set to 1 pixel. You can get rid of the border by
setting it equal to zero.
FIGURE 7.17
The structure of
a table.
Tech Talk
though there is a cascade in internal and inline styles—a background color you
set inline in the body will override one you set in the head—when people refer
to cascading style sheets, they generally mean external style sheets. If you have a
large site and want to make a change to all your pages, you only need to change
one piece of code in one CSS file. You might have a 100-page site, but you can
change the color of all of your paragraphs with one small change to your style
sheet. Much like regular HTML, most good web design tools like Dreamweaver
will help you create your styles without getting too heavily into the code. Also
like HTLM, CSS has different versions. Although the W3C is working on CSS
3, CSS 1 is the most widely supported version.
External style sheets use selectors to tell the browser what characteristics to
apply the rule to. For instance, you might choose to make all first-level head-
ings (<h1>s) blue. In your style sheet, you would write a bit of code (essen-
tially a short sentence) to change all of those headings at once. In that case,
h1 is your selector. You can select most elements and control their appearance
and sometimes behavior in CSS. In addition, if you have a special element on
your page that you want to treat differently, you could create an id selector that
applies only to one element on a page and format it differently. In the HTML,
you’d label the code using the id attribute. Perhaps you want to control your
navigation div’s position, fonts, and so on from your external style sheet (<div
id=“nav”></div>). Perhaps you want the button with the id “thispage” (the page
you’re on) to be a different color than all of your other navigation so the site’s
menu can indicate what page the user is on. Or you could create a class selector
that can apply to several different tags, perhaps bits of text you want to put boxes
around (<p class=“boxed”></p>).
When you are debating whether to use a class or id when writing your code,
think about how the section will be used. If it is a section of the page that is
only going to be used once per page, such as a page footer, it should be an id.
Otherwise, set it up as a class. Having trouble keeping track of this? Here’s a
helpful mnemonic—a student may have many classes, but should have only one
student id card.
CHARACTERISTICS OF APPEARANCE
You have a basic structure and content. The properties that follow (using CSS)
let you control the look. These characteristics can be controlled by inline styles
or internal style sheets, but they are best controlled by external style sheets.
Remem- ber, separate the meaning of your content from the presentation to make
the code easier to understand (somebody needs to maintain the pages), to make
sitewide changes easier, and to make the content more accessible to users with
disabilities.
Margins
Web page real estate is valuable. Although large margins are good on paper,
most designers like to have small margins. When you set top and bottom
margins,
Web CHAPTER 7 213
Design
you are setting the space above and below your text and images. This space
is added to the very top and bottom of your page. It is not a buffer of space,
so it disappears when you scroll down. Left/right margins work in a similar
way.
Define the value of your margins using the number of pixels (px) or one of
sev- eral other measurement units (for instance, % for percent of the page
width or height; pt for points; mm for millimeters; or em for em widths, the
width of an “m” in whatever font size you use or whatever font the user’s
browser displays). You can define top and bottom margins, and you can also
define left and right margins.
Tech Talk
ABSOLUTE SIZES
One way to set your font size in HTML is by using absolute values. Instead
of using points (as you would in a word processor), you’ll use numbers, with
1 being the smallest font and 7 being the largest. If you don’t set a font size,
your paragraph text will be displayed in size 3 font (16 point font).
RELATIVE SIZES
Another way to set your font size in HTML is by using relative values. To
use relative values, you first need to set your base font size (reset the default
font size in the page properties window). Then, you select the relative sizes 11
to 16 for larger fonts and 21 to 26 for smaller fonts. The difference in font
size will always be relative to the base font size you established for the site.
Link Characteristics
How do users know what text on a web page is a link or which links they
have already followed? A site is more usable if users can identify links and tell
which links they clicked on. Designers set different characteristics for links to
make navigating easier using pseudo-classes, such as a:hover, which changes
the properties of a link when the cursor is over the link or a: visited, which
changes the color of links to already visited pages. Although links are often to
web pages or other files, you can also use them to take users to a specific part
of a web page.
If you don’t want your linked text to be blue and underlined, you can change
how it appears, but make sure you give the user good cues that the text is linked.
Create your links in a distinct color, and style them so the browser adds under-
lining when you move the cursor over them (mouse over or hover). You can be
cre- ative, but apply color theory (such as by choosing colors for effect and
selecting color analogs for your link pseudo-classes), and be consistent with link
colors throughout your site.
To control text link appearance and states (how the link appears before, dur-
ing, and after the user interacts with it), use an external style sheet. In your style
Web CHAPTER 7 215
Design
sheet, you can control these and other states with pseudo-classes (a selector
followed by information about the element’s state):
■ Unvisited. This is how your link shows up before you click on it. It should
stand out from the page. Use the selector a:link.
■ Hover. This is how your link appears when you mouse over, or move the cur-
sor over, the link. Use the selector a:hover.
■ Active. This is how your link shows up when you are clicking on the link and
haven’t yet released the mouse button. Use the selector a:active.
■ Visited. This is how your link shows up after you have visited that page or
document. Use the selector a:visited.
Like any text, you can control all sorts of properties of your links: margins,
background colors, background images, borders, text decoration (like under-
lining), and so on, creating text-based buttons styled in CSS. The key is set-
ting properties for separate states. Figure 7.18 shows how you can change
link colors and make the underlining appear only when you move the cursor
over the links
FIGURE 7.18
Controlling link
pseudo-classes
in CSS.
Tech Talk
Accessibility Issues
We’ve talked about accessibility throughout this chapter and in Chapter 6,
“Interface Design and Usability.” When you think about your audience, always
remember that some of your users may have different needs because of dis-
abilities. The most common issue that you’ll hear about is users who have a
vision problem. These problems can range from color blindness to problems
reading small print, to users who are completely blind and rely on a text reader
to use the Web. Other disabilities you should consider are hearing loss; motor
218 SECTION 2 Multimedia Design
control issues (users who can’t use a mouse, for instance); and cognitive prob-
lems, including, among others, problems with attention, memory, and reading.
As you build your site, make sure it’s usable for users with disabilities. Here are
a few questions you should ask to start off:
■ If I took away all of the nice presentation and was left with only text,
links, and media …
■ could I access the information?
will see your page. Mozilla’s FANGS add-on will take an open page in Firefox
and simulate how a text reader would render it. FANGS will let you see what
these users will likely get if they pull up a headings list or a links list, too. Use
these tools, but rely on common sense as well. Electronically detectable prob-
lems are a start, but you need to develop pages that make sense, pages that are
usable by users with disabilities. Accessibility specialist Neal Ewers uses a badly
designed form as an example: if a blind user has to read through an entire form
before reading the context and instructions below it, the form isn’t very usable.
Making your page accessible for any of these users will frequently increase
the overall usability of the site on both computers and mobile devices. And
as you learned in Chapter 6, “Interface Design and Usability,” you may have
a legal obligation to make your site accessible, too.
As the tools and technology for creating and uploading web pages continue to become
CHAPTER SUMMARY
more user friendly, remember that production software and hardware issues play a
relatively small role in determining the effectiveness of a website. The best sites
will always be those with good navigational structures that are functional, easy to use,
and accessible to all audiences; that provide substantive, timely, and meaningful
content for the user; that feature attractive designs and page layouts; and that are
maintained and updated by the webmaster on a regular basis. Such a site is only
possible with thoughtful research, planning, design, and authoring on the part of the
webmaster or production team.
220 SECTION 2 Multimedia Design
SECTION 3
Static Media
8. Graphics 217
9. Text 245
10. Photography 277
Reeed/Shutterstock.com
This page intentionally left blank
CHAPTER 8
Graphics
217
The whole is greater than the sum of its parts. Key Terms
—Origins Unknown (but often attributed to Additive Color Mixing Aliasing
Gestalt Psychology, 1920s) Anti-aliasing Aspect Ratio ATSC
Bit Depth CMYK
Color Space Compression Cropping
DVB EDTV
Chapter Highlights Field Frame Frame Rate Graphic H
This chapter examines: Hertz Image
The nature of computer graphics Interlaced Scanning JPEG
Moving Image Native Resolution N
The raster image encoding process
Optimization PAL
The vector graphics encoding process
Path
Moving image scanning technologies Persistence of Vision Pixel
Computer and television display technologies Pixel Count
World standards for Digital Television (DTV) broadcasting Pixel Dimensions Progressive Scan
(or Bitmap) Refresh Rate
Graphics
A graphic is any type of visual presentation that can be displayed on a
physical surface like a sheet of paper, wall, poster, blackboard, or computer
monitor. Graphics are a product of human imagination and are typically
created by hand or with computer-assisted drawing and design tools.
Graphics include things like stick figures, symbols, numbers, drawings,
typography, logos, web buttons, illustrations, and line art (see Figure 8.1). A
graphic designer is a media arts professional who creates graphics for use in
print or electronic media.
FIGURE 8.1
This assortment of
graphics includes
clipart, logos, line art,
and symbols.
Images
An image is a two- or three-dimensional representation of a person, animal,
object, or scene in the natural world (see Figure 8.2). Images can be still or
moving. A still or static image is one that is fixed in time. A moving image—or
time-based image— is one that changes over time. Photographs, maps, charts,
and graphs typically fall into the still image category. Broadcast television,
digital video, and motion pic- tures are examples of moving images. As we’ll
see later in this chapter, moving images are made by presenting a sequence of
still images in rapid succession to simulate the illusion of motion. In reality,
there are no moving images, only the optical illusion of movement created by
the repetitive presentation of static images.
DIGITAL IMAGING
A film camera uses a plastic strip, coated with a light-sensitive emulsion, to
record a scene composed by the photographer. The film negative that’s produced
is real and can be handled, held up to the light, or passed along “physically” to
someone else. In the digital world, everything is reduced to a number, including
graphics. For example, a digital camera uses an optical image sensor to con-
vert light into electrons (electrical energy). The electrical signal is converted into
a digital recording and saved as a binary file made up of 0s and 1s. While a
binary file cannot be touched or held up to the light, it is every bit as real to the
computer as the film negative is to the photographer (see Figure 8.3).
Graphics CHAPTER 8 219
FIGURE 8.2
A large collection of
photographic images.
FIGURE 8.3
Brothers examine a strip of photographic negatives. (Left) In analog photography, the negative
is used to make photographic prints. Because digital images are recorded numerically, they
cannot be directly touched or viewed. (Right) A digital device or computer is required to render a
binary image for output to a display screen or a printer.
220 SECTION 3 Static Media
Two methods are commonly used to digitally encode and display computer
graphics. The first approach, called bitmap or raster imaging, uses pixels to
define the structure of a digital image. Tiny squares of color, like tiles in a
mosaic, make up the graphic. Depending on the number of pixels, or squares,
per inch, you may not even notice them in the final digital or printed graphic.
The second approach, called vector imaging, uses mathematically constructed
paths to define a graphic’s visual structure. In other words, it records a
graphic as a group of interrelated points, lines, curves, and shapes. Table 8.1
compares some of the differences between the two methods.
Table 8.1A Comparison of the Raster Image and the Vector Graphic
Scalability Resolution-dependent
(fixed number of pixels), Resolution-independent, so
so image quality graphic can be resized
deteriorates when without losing detail or
enlarged. clarity.
Common File
Formats .bmp, .gif, .jpg, .png, .eps, .svg, .swf, and
and .tif .wmf
File Size Typically large. Raster images
can be compressed to reduce Relatively small.
file size. Some compression Vector encoding is
formats may cause the loss of highly efficient.
image details.
Graphics CHAPTER 8 221
FIGURE 8.4
Thirty-five red pixels coalesce to form this bitmap graphic of a capital letter H.
Raster Images
A raster image is formed by dividing the area of an image into a rectangular
matrix of rows and columns comprised of pixels (see Figure 8.4). A pixel, short
for picture element, is a square (or occasionally rectangular) area of light
repre- senting a single point in a raster image. Every pixel in a raster image is
exactly the same size and contains a single color value that’s typically stored as
a 24-bit string of binary data. The total number of pixels in a raster image is
fixed. In order to make a raster image physically larger, more pixels have to
be added to the raster matrix. Likewise, pixels need to be discarded when
making a raster image smaller. The width and height of a raster image is
determined by how many pixels each row and column contains.
On their own, pixels are relatively meaningless, but when combined with
hundreds, thousands, and even millions of other pixels, complex patterns and
photorealistic images can be formed. German psychologist Max Wertheimer
(1880–1943) developed the concept of perceptual grouping to explain the
human tendency to perceive whole shapes and patterns from an arrangement
of smaller particles of visual information. This concept is commonly
expressed as “the whole is greater than the sum of its parts.” Take a look at
the image in Figure 8.5 to see perceptual grouping at work.
The mosaic façade in this photograph is made up of thousands of
individually colored tiles carefully arranged to form an exquisitely detailed
composite image. It takes very little effort on the part of the viewer to overlook
the individual pieces
222 SECTION 3 Static Media
FIGURE 8.5
From a distance,
the individual pixels
forming this image
are barely perceptible
to the naked eye. In
fact, our brain works
hard to achieve and
maintain a holistic
impression. Up close,
however, we can see
that many small pieces
of visual information
went into forming this
19th-century mosaic
of Christ, the good
shepherd.
Source: Historien d’art
(Own work) [Public
domain], via Wikimedia
Commons
FIGURE 8.6
Digital image pixels
are much smaller
than the bits of tile
used in mosaic art.
You have to zoom in
really close on this
image before you’ll
see the pixels used to
form it.
of glass and stone used by the artist. Instead, we’re much more inclined to
perceive the scene holistically, forming the impression intended by the artists’
careful and purposeful arrangement. This technique of using tiny bits of
colored material to form a composite visual impression dates back to about
3,000 BC and is still used today in the print and electronic media industries to
convey visual information. In Figure 8.6 we see a digital photograph of a fruit
Graphics CHAPTER 8 223
basket. This
image too is
constructed
224 SECTION 3 Static Media
of individual colored tiles. Millions of them in fact! These tiny pixels can be seen
by zooming in on the image using a photo editing program such as Adobe
Photoshop.
RESOLUTION
Resolution describes the image quality of a raster image and directly relates to the
size and quantity of the pixels the image contains. In the illustration in Figure
8.7, the first drawing of the triangle has only three dots, making it a low-
resolution image.
As more picture elements are added, the quality of the image improves consider-
ably, moving it along a continuum from low- to high-resolution.
Simply put, the more pixels you have in a given area (for example, in a
square inch), the more information you have, and the higher the resolution of
the image. In the example in Figure 8.8, artist Peter Roche used nearly
10,000 Jelly Belly jellybeans to form this portrait of President Ronald Reagan
(left). By com- parison, the official White House portrait of President Reagan FIGURE 8.7
(right) consists of more than 4 million pixels. This sequence of
four triangle graphics
Jellybeans are a creative medium for artistic expression, but they are quite large progresses from
compared to the pixels used to from a digital image. Because of this, the low-resolution to high-
image detail in the jellybean artwork pales in comparison to the resolution of resolution (left to right)
as visual information
the actual photograph. The pixels in the digital photo are so small that they
and detail are
are undetectable with the naked eye. added. Interestingly,
it only takes three
dots to trick the
eye into perceiving
the shape of a
triangle. Remember
the principle of
psychological closure
which we discussed in
Chapter 4?
FIGURE 8.8
The first portrait of
President Ronald
Reagan (left) was
commissioned by
the Jelly Belly Candy
Company. It was later
donated to the Ronald
Reagan Presidential
Library in Simi Valley,
California.
(Courtesy of the Ronald
Reagan Presidential
Foundation and the Jelly
Belly Candy Company.
Artist: Peter Roche.)
Graphics CHAPTER 8 225
Tech Talk
FIGURE 8.10
The Adobe Photoshop CS5 color picker is used for mixing
colors. Each color shown here was achieved by changing
FIGURE 8.9 the values in the red, green, and blue property fields.The
The primary and secondary colors of white light become hexadecimal code is shown beneath the RGB property
visible when light is refracted through a glass prism. fields and is used to quickly identify or pull up a particular
226 SECTION 3 Static Media
color.
Graphics CHAPTER 8 227
process, each page goes through four presses, the bit depth, or number of bits the device uses
each inked with a primary color or black pig- to record intensity and color components. And
ment. The letter K refers to the black and the greater the bit depth, the larger the color pal-
comes from the term key plate, a black printing ette the image will have. Bit depth gets confus-
plate. If you are designing for print, you will ing, however: when we talk about 8-bit graphics,
likely create CMYK images. The challenge, of we sometimes mean 8 bits total (256 colors),
course, is that you’ll be creating them on an but at other times we mean per channel (per
RGB color monitor! pri- mary color), or 24 bits (16.8 million colors)
(see Figure 8.12).
Bit Depth or Color Depth Some display systems can render an even
When you capture an image with a digital higher bit depth (up to 48 bits), but the 24-bit
cam- era, camcorder, or scanner, the device color stan- dard is currently the most common
encodes light into electrical energy, and then into and will give you a sufficiently large palette
bits for storage in a format that the computer for multimedia applications. You may also
can pro- cess and understand. Display devices hear about another channel, the alpha
reverse the process by decoding bits back into channel. This is an added 8-bit channel that
electrical energy and light impulses that can records transparency informa- tion in some
be rendered on screen. The more data you image formats.
collect, the greater
FIGURE 8.11
RGB color space (left) is used for
multimedia (Web, animation, and
television, etc.). CMYK color space
(right) is most often associated with the
four-color printing process.
FIGURE 8.12
The possible color combinations for
any pixel in an 8-bit graphic (or a
24-bit display). If you follow the
arrows through every possible data
combination, you’ll get 256 (or 28)
possibilities for each color channel.
Combining channels—256 possibilities
for red 256 for green 256 for
blue, or 224 combinations—you’d
have about 16.8 million
possible combinations.
Illustrator: Susan A. Youngblood.
228 SECTION 3 Static Media
Pixel Dimensions
When we talk about pixel dimensions, we’re not talking about the size of an
individual pixel. Instead, we use the term pixel dimensions to describe the size
of a raster image, expressed as the number of pixels along the x-axis (width) by
the number of pixels along the y-axis (height). For example, an 800 600 pixel
image contains 800 pixels across the image from left to right and 600 pixels
across the image from top to bottom.
Pixel Count
Pixel count is the total number of pixels in a raster matrix. To determine the
pixel count, multiply the horizontal and vertical pixel dimensions. The 30
18 pixel image in Figure 8.13 has a pixel count of 540 pixels.
FIGURE 8.13
Pixel count is
determined by
multiplying the number
of pixels across a
digital image by the
number of pixels high.
Graphics CHAPTER 8 229
FIGURE 8.14
This chart displays
the resolution sizes
and compression
settings for the Canon
G12 digital camera.
this point, let’s consider the Canon G12 digital camera, which has an
effective resolution of about ten million pixels (total pixel count) (see Figure
8.14). With such a large-capacity image sensor, the G12 can produce a
photograph with a recorded pixel count of 3,648 2,736 pixels. That’s a lot
of pixels! Dividing both pixel dimensions by 300 allows you to determine the
maximum size of a photographic print that can be made from this image with
good results.
3,648 pixels 300 pixels/inch 12.16 inches.
2,736 pixels 300 pixels/inch 9.12 inches.
A photographer won’t always need to produce a print this large, but having
lots of pixels to work with is always better than not having enough.
In multimedia work, we’re much more concerned with display resolution and
bandwidth than we are with print resolution. Most television and computer
monitors have a display resolution of either 72 or 96 ppi (that’s a pixel count
of 5,184 or 9,216 in a square inch). So images produced for the screen need
somewhere between a seventeenth and a tenth as many pixels, respectively, as
those produced for print in order to look good. For video and the Web, 72 ppi is
the industry standard; on a 72 ppi monitor, each pixel in a 72 ppi image will be
displayed by one pixel on the screen. You can go as high as 96 ppi, but anything
more than this is simply a waste of file bandwidth and will not increase the
overall quality of an image that’s displayed electronically.
SCALING
Many software applications allow you to scale an image within an open
docu- ment by selecting it and adjusting one of eight resizing handles along
the outer edge. But raster images are resolution-dependent, which means they
contain a fixed number of pixels. Resizing (or scaling) a raster image without
redefining the structure and pixel count of the array (resampling) can ruin
your image.
When you resize this way, you don’t change the image matrix (the image’s pixel
dimensions) or the amount of data stored. You only decrease or increase the
size of your pixels. When you scale an image upward (make it larger), each pixel
is enlarged, and you lose image detail and sharpness. The more you enlarge a
raster image, the softer and fuzzier it becomes. For this reason, professionals try
to avoid the enlarging of raster images (see Figure 8.15).
230 SECTION 3 Static Media
FIGURE 8.15
Scaling is the act
of resizing a digital
image to make it
appear smaller or
larger on screen.
Photographer: Sarah
Beth Costello
RESAMPLING
Resampling changes the size of a raster image by increasing or decreasing
the image’s pixel count. While on the surface this sounds like a simple process,
you must remember that each pixel represents a single color value. If you add
pixels to an already defined image, what color do you to assign to them, and
where do you place them? Which pixels get shifted to make room for the
new ones? Likewise, if you delete pixels from an image to make it smaller,
which ones get tossed and which ones get to stay? Resampling deals with
these challenges by using algorithms to analyze each pixel’s color
information and using this data to reconstruct an entirely new raster structure.
Depending on which resam- pling method and algorithm you use, some of
the original image data may be retained, but much of it may be discarded and
replaced. For this reason, you should make a backup copy of your original
image before applying changes.
When you resample to enlarge an image, you still lose detail and sharpness.
Given the nature of raster images, this just can’t be avoided. However,
resampling provides more options and typically yields better results than
scaling alone.
ANTI-ALIASING
Raster images are also known for producing aliasing artifacts, the visibly
jaggeddis- tortions along the edge of a line. Aliasing is a stair-step effect caused
by using square pixels to define objects with curves or diagonal lines (see Figure
Graphics CHAPTER 8 231
8.17).
You
can
easily
see
the
effect
when
looki
ng at
text
on the
scree
n of a
small
digita
l
devic
e such
as a
cell
phone
.
232 SECTION 3 Static Media
(a)
(b) (c)
FIGURE 8.16
(a)Scaling. Upscaling often results in a noticeable loss of image quality (increased blurriness). When downscaling a high-
resolution image, image degradation is rarely a concern; (b) Resampling. The original image was too big to fit on this
page. I used Adobe Photoshop CS5 to resize (and resample) it to the version you see printed here. Photoshop offers you a
choice of five resampling algorithms. Because this image was intended for print, I kept the resolution set to 300 ppi. If I
wanted
to publish it to the Web, I would have chosen 72 ppi. To reduce the loss of image quality during scaling, be sure to select a
resampling method (Source: Neale Cousland/shutterstock.com); (c) Cropping. The two images on the right were achieved by
cropping the original photo (top left ). Cropping is a photo editing technique used to delete portions of an image in order to
enhance the focus of a main subject or improve composition. With scaling, the original composition is preserved by physically
resampling the image using fewer pixels (downscaling). With cropping, pixels in the unwanted portion of the image are
permanently deleted. The remaining pixels are preserved with their original color values intact. Resampling is not performed.
A cropped image will always, by definition, be smaller than the original; however, this reduction in size is due to the deletion
of image content (pixels) and not to scaling.
Graphics CHAPTER 8 233
Anti-aliasing smooths out the edges of jagged type by blending the color
transition points, such as the pixels along the edges of a letter (see Figure 8.18).
The only major drawback to this is that it increases file size somewhat. In most
cases, it’s better to have a clean graphic and accept the slightly larger file size.
Anti-aliasing typically works best on larger type as the jagged edges of the type
are more visible.
Tech Talk
Compression In an uncompressed file format, the
drives were much smaller, and processor speeds weren’t
graphic file records the individual value of each pixel.
nearly as fast as they are today. In addition, data transfer
Examples of this include the BMP and uncompressed
rates were particularly slow online—a fast modem at the
TIFF formats. While these formats give you access to
time was around 56kpbs. Bits and bytes were
a lot of information, and TIFF in particular is a
precious, and the new JPEG standard greatly improved
popular format for photographers, graphics saved in this
the photo- imaging workflow. When applied to a raster
format tend to be quite large, which is not always a good
image, the JPEG compression algorithm evaluates each
choice for online projects or storage. Compression can
pixel, looking for ways to “compress” redundant color
help with this. There are two basic types of compression:
information into a more efficiently written and structured
lossless, which looks for more efficient ways to store
data file. For exam- ple, the high-resolution photo in
the data without losing any information—kind of like
Figure 8.19 was taken on a clear and sunny day and
putting your sleeping bag in a stuff sack—and lossy,
contains a lot of blue pixels.
which while re- ducing the file size gets rid of data
that you might not need at the moment. The original image size is 4,288 2,848 pixels, but
notice how the first 100 rows of this image contain
JPEG is the most common lossy format used in multi-
largely the same shade of blue. We can compute a
media production. Released in 1992 by the Joint
rough estimate of the uncompressed file size of this
Photo- graphic Experts Group, the JPEG standard was
sample as follows:
designed to reduce the file size of photographic
images. File size was a critical issue at the time Pixels Count: 4,288 pixels per row 100
because computer hard row
248,800 pixels
234 SECTION 3 Static Media
File Size: 428,800 pixels 24 bits
10,291,200 bits
Graphics CHAPTER 8 235
FIGURE 8.19
This high-resolution
photograph contains
lots of redundant color
information (blue sky
and golden wheat)
making it a great
candidate for JPEG
compression.
FIGURE 8.20
Using Adobe
Photoshop’s Save
for the Web interface
allows you to perform
a side-by-side
comparison of the look
and size of the original
image on the left, with
a preview image on
the right. Every time
you make a change
to the compression
settings, the preview
image is automatically
updated to reflect that
change. The goal of
image optimization
is to produce the
highest-quality image
with the smallest
file size and no
236 SECTION 3 Static Media
compression artifacts.
Graphics CHAPTER 8 237
FIGURE 8.21
Another example of compre ssion at work. At 50%, can you notice any degradation?
Which parts of the image are most highly affected by compression?
solid blocks of color. It supports interlacing, so every odd line of pixels loads,
then every even line loads, making graphics seem to appear faster (users see
the full-sized, half-loaded graphic before the rest of the pixels appear).
■ JPEG offers 16.8 million colors but does not support transparency (has no
transparent pixels). It is a lossy compression format and is used most
often for photographs. This format does not support interlacing.
■ PNG offers 16.8 million colors and transparency, but you can choose to
use fewer colors to save file space (PNG 8, or PNG with 8-bit color). It is a
lossless compression format and is common for a wide range of images,
including favicons (the small web page icons in browser tabs). Some older
web brows- ers don’t support it (Internet Explorer prior to version 4); such
browsers have mostly, but not completely, fallen out of use. PNG files can
be very small, but for photographs with many colors, they may be larger
than comparable JPEGs. This format supports interlacing.
Another option you have with both the GIF and PNG formats is dithering, or
scattering the pixels to achieve blending without using as many colors.
Dither- ing is useful if you have an image with a drop shadow and want to
superimpose it cleanly on a background.
Vector Graphics
Vector imaging defines the area of a picture using paths made up of points, lines,
curves and shapes. Each vector path forms the outline of a geometric region con-
taining color information. Because paths can be mathematically resized,
vector graphics can be scaled up or down without losing any picture clarity.
Clipart and typefaces (fonts) are often created and stored as vector graphics
because designers want the ability to scale them to any size (see Figure 8.22).
FIGURE 8.22
Vector graphics have
crisp edges with
no aliasing. They
can be resized up
or down to any
size without
negative
consequences.
Graphics CHAPTER 8 239
FIGURE 8.23
To complete a paint-by-numbers piece like this one, each numbered region must be filled in
with a single color. A vector graphic is rendered geometrically in much the same way using
paths defined by points, lines, curves, and shapes. (Courtesy of Pam Snow, Mesa, AZ )
Theconceptbehindvectorgraphicsislikepaintingwithnumbers. Apaint-by-numbers
set normally includes a black and white line drawing and a set of numbered
paints. The artist fills in each numbered region with the appropriate color,
carefully staying within the bordered outline of each defined area. When all of
the regions are filled with color, the picture is complete. As with raster images,
the phenomenon of per- ceptual grouping leads us to ignore the individual
paths used to form the holistic impression (see Figure 8.23).
Vector graphics can render curves and diagonal lines that are crisp, smooth, and
sharp. Aliasing is not a problem because pixels are not used in their construc-
tion. So vector graphics are an ideal choice for prepress applications
requiring higher-resolution pictures with finer line detail.
When you enlarge a raster image, the file size grows in proportion to the size
of the image: as you add pixels to the array, you need more data to represent
the image. Because vector encoding uses mathematical equations to record
visual information, the size of a vector data file stays consistent, regardless of
how large or small you make the graphic. If you are creating a still graphic,
you can enlarge the graphic to any size you want, then rasterize it, saving it to
whichever file format suits your purpose best.
240 SECTION 3 Static Media
You could also use vector graphics to create an animation, such as with
Flash. Instead of drawing every separate frame of your project—with 24 frames
appearing each second—you could create two different graphics for a segment
and let your animation software mathematically interpolate the positions of
the components in the in-between frames (a technique known as tweening).
FIGURE 8.24
Graphics CHAPTER 8 241
A comparison of digital screen sizes and resolutions for a handful of mobile devices.
242 SECTION 3 Static Media
(the image becomes fuzzier as you stray further from the native resolution
of the monitor).
2. A user may not have the screen set to its native resolution but may be zoomed
in on an active document window. For example, you could be viewing an
online newspaper article using a web browser like Firefox, Chrome, Internet
Explorer, or Safari. Most browsers allow you to zoom in on a page to get a
better view of the content. Doing so enlarges the view of both text and
images, however, with each increase you’ll lose clarity, particularly with
images (see Figure 8.25).
FIGURE 8.25
A web page is viewed natively in its actual size (top). The close-up view (bottom) was achieved
using the browser’s zoom control.
Source: www.sptimes.ru
Graphics CHAPTER 8 243
FIGURE 8.26
A comparison of some common display resolutions used in multimedia design. Do you notice
how the aspect ratio varies?
Aspect Ratio
In addition to describing the screen attributes in absolute terms (screen size
and native resolution), monitors are classified by their aspect ratio. Aspect ratio
is an indicator of the proportional relationship of the width to the height of
the screen and is depicted with the expression x:y, where x equals the number
of units wide and y equals the number of units high. While the physical size
of a display screen can vary, the aspect ratio remains constant. The two most
common aspect ratios in use today are 4:3 and 16:9. The standard 4:3 (pro-
nounced 4 by 3) aspect ratio predates television and produces a familiar and
somewhat boxy-looking shape. The other popular aspect ratio is 16:9 and is
usually referred to as widescreen because it more closely matches the shape of a
theatrical movie screen. While television and computer monitors are available
in many shapes and sizes, they almost always conform to either a 4:3 or 16:9
aspect ratio.
MOVING IMAGES
Many people’s daily experience is filled with moving images. Some come from
televisions and movie theaters. Others come from personal computers, game sys-
tems, mobile phones, handheld devices, even GPS interfaces and self-checkout
244 SECTION 3 Static Media
kiosks at supermarkets. Regardless of the content, they are typically based on the
same basic principles. Let’s look at how this technology works.
Raster Scanning
In the Western world, people tend to process visual information from left to
right and from top to bottom. Think about it! When you write a letter or note,
you generally start in the upper left-hand corner of the page and work your
way down one line at a time from left to right. On a typewriter, you do
basically the same thing. You press a key to produce an imprint of a
character on the page, and the writing head advances to the next space on the
line. You advance one character at a time until the end of a line, then the
writing head shifts leftward and downward to the beginning of the next line
of text. A sheet of paper has a fixed number of lines on it, so when you reach
the end of a page, you need to load in another one to continue.
The process of raster scanning works in much the same way, only faster. In
television and computer display systems, individual video frames and computer
images are reproduced on the screen, one pixel at a time, in a process called
scanning. An electron beam or impulse mechanism illuminates each screen pixel
as it progresses through the raster matrix. Each row of pixels is called a scan
line. A scanning cycle is one complete pass of all of the scan lines in the display.
When the scanning beam reaches the last pixel on the last scan line, it moves
to the top and begins the next cycle. A frame is one complete scanning pass of
all of the lines in a picture, or one complete scanning cycle.
The refresh rate is the number of complete scanning cycles per second and is
measured in Hertz (Hz), a unit of frequency equal to one cycle per second. If the
refresh rate is below 50 Hz, the image will appear to flicker. Most displays have
refresh rates of 60 Hz or more. The faster the refresh rate, the sharper the image
quality will be and the less eyestrain the user will experience. The larger the
screen, the higher the refresh rate should be. Large computer monitors typically
have a refresh rate of 85 Hz or higher.
PROGRESSIVE SCANNING
Contemporary computer monitors and some televisions reproduce images using
progressive scanning, consecutively scanning the lines of the picture from top to
bottom, just as you type on a typewriter. Progressive scanning helps combat
eyestrain, which is why it’s a given on computer monitors. That’s not,
however, necessarily the case for television.
INTERLACED SCANNING
Early television standards adopted a method of raster scanning called interlaced
scanning to minimize both bandwidth use and flickering. With an interlace
sys- tem, each frame of an image is captured in two parts and transmitted
separately, one field at a time. The odd lines are scanned first, followed by a
second pass of the even lines. So you’re really only seeing half of each new
image at once, but the screen draws so quickly you don’t notice.
Graphics CHAPTER 8 245
FIGURE 8.27
Broadcast television images are typically interlaced (left) while video on the Web is often
de-interlaced (right), delivered progressively.
Fields
One complete scanning pass of either the odd or even scan lines is called a
field. So two fields, the odd and even, produce one frame. As you can imagine,
the electronic raster scanning process has to be fast to give you a good
picture. Let’s say you’re watching a movie on a television with interlacing
and a frame rate of 30 frames per second (usually stated as 30 fps) that has
480 lines in its raster (a comparatively small number). This means that one
scanning pass of the 240 odd-numbered scan lines, or one field, occurs in
just 1/60th of a second. Double that to get a full frame. Put another way,
14,400 scan lines of picture information are rendered on your television
monitor every second.
Television Standards
Since multimedia projects are often viewed on televisions, you need to con-
sider television standards. A great deal of money goes into supporting the
infrastructure of terrestrial broadcasting systems, so countries have developed
technical standards—some more widely adopted than others—for the produc-
tion of television-related equipment. While it doesn’t always work out this way,
such standards help to ensur e that consumers have access to equipment that’s
compatible with the delivery systems used by content providers for program
246 SECTION 3 Static Media
GREAT IDEAS
The Illusion of Apparent Movement
The foundation of all moving image technology rests on the ability of the human
eye and the brain to process a series of rapidly projected frames or scan lines as
a
continuous and uninterrupted picture. The motion we observe on the screen, whether
in a movie house or on a television or computer monitor, is a perceptual illusion. Film,
video, and moving digital images appear to have motion because of the phenomenon
of short-range apparent motion. Our brains process successive images that have
small changes between them as movement, in much the same way that they process
real motion.
In order to pull off the illusion of motion we get from film, video, and animation, indi-
vidual pictures in a sequence must advance quickly. If the frame rate is set too low,
the transition from one image to the next will appear jerky or stilted. The target speed
is known as the flicker fusion threshold, the frequency at which the momentary flicker
of intermittent light between each frame disappears from human perception.
Early on, the motion picture film industry adopted a frame rate of 24 fps as an
international standard. However, an image pulsating at 24 fps is well below the flicker
fusion threshold for human perception. To compensate, a projector displays each
frame of motion picture film twice. A rotating shutter momentarily blocks out the
projector’s light each time the frame is advanced and between each repeated exposure
of a single frame, fixing the image in one spot and keeping us from seeing a blur. We
don’t notice the brief black spots because of persistence of vision, the phenomenon in
which our retinas hold a visual impression of light for a fraction of a second. The result
is a flicker-free viewing experience for the audience.
FIGURE 8.28
Moving images are a perceptual illusion, achieved by the rapid projection of individual still
frames of film or video.
Graphics CHAPTER 8 247
distribution. In an ideal world, every nation would use the same standards for
every type of electronic technology, but this just isn’t the case. As it’s sometimes
hard to get two people to agree on something, it’s even more difficult to get
the governing bodies of entire nations to agree on a universal set of technical
specifications.
As you develop media products to be used on multiple devices—computers,
phones, game systems, and so on—keep in mind how those products will work
and look on each device, including television.
FLASHBACK
The Legacy of Analog Television
Television signals used to be broadcast in analog formats. In analog broadcasts, continuous waves carried the sound and pictu
Much of the world has made the switch to digital formats, and many (but not all) remaining countries have plans to switch t
Each analog format had a 4:3 aspect ratio, set to mimic the dimensions of the film that was in use when television was born.
With the release of high-definition digital television, the aspect ratio was changed to the now popular 16:9 widescreen format, b
FIGURE 8.29
Television
entertainment
technologies have
evolved rapidly in
recent years. The
standalone single-
piece television
receiver your parents
may remember
can’t compete with
today’s high-tech
home theater system,
complete with wall-
mounted flat-screen
monitor and 5.1
surround sound.
248 SECTION 3 Static Media
DIGITAL TELEVISION
Digital television (DTV) offers many advantages over legacy analog formats.
Content created for digital media is more fluid: it can be easily repurposed
and distributed through secondary channels of communication, making DTV
more compatible with computer and Internet-based systems and services.
DTV also offers less signal interference and uses less bandwidth than an equiva-
lent analog television broadcast, which is an advantage because the amount
of broadcast bandwidth is finite. The switch to DTV has meant that more sta-
tions can be broadcast in the same viewing area, while using the same or less
bandwidth as analog television. DTV also offers the option of using a 16:9
for- mat, similar to that used in the movie theater, as well as high-definition
(HD) video—video with over twice the resolution of the old NTSC standard.
When professionals shoot and edit television programs digitally, the DTV
infrastruc- ture preserves the quality of the original material during
transmission. In order to transmit digital content through an analog system,
programs must first be downconverted to an analog format, resulting in a loss
of image quality.
ATSC
The United States adopted the ATSC (Advanced Television Research Consor-
tium) terrestrial broadcasting standard in 1996. In the same year, WRAL in
Raleigh, North Carolina, became the first television station in the country to
begin broadcasting a high-definition television signal. The U.S. transition to
HDTV was fraught with many delays and took more than a decade to complete.
On June 12, 2009, U.S. analog transmissions ceased and NTSC broadcasting
officially ended in the United States.
The NTSC format has a fixed resolution, aspect ratio, scan mode, and frame rate.
The newer ATSC standard is more fluid, providing up to 18 different display
formats, which are categorized into three groups: standard definition
television (SDTV), enhanced definition television (EDTV), and high-definition
television (HDTV) (see Table 8.2). ATSC emphasizes progressive scanning
and square pixels, bringing television technology closer to current standards
for computer imaging. It also improves audio distribution, enabling a theater-
style experi- ence with 5.1-channel Dolby Digital Surround Sound. The
ATSC standard has been adopted in much of the Americas and in U.S.
territories. Canada made the switch in 2011, and Mexico is preparing for the
switch and is simulcasting in both digital and analog formats. Other countries
also have introduced the ATSC format but have not fully switched. And, of
course, the ATSC is working on new standards: as ATSC 2.0 comes out, look
for features such as video on demand and possibly even 3D programming.
DVB
The DVB (Digital Video Broadcasting) terrestrial broadcasting standard was
established in 1997. The following year, the first commercial DVB broadcast
was transmitted in the United Kingdom. Because European consumers depend
more on cable and satellite distribution for television and less on over-the-air
Graphics CHAPTER 8 249
245
Yes, it is a press, certainly, but a press from which shall flow in Key Terms Alignment All Caps As
inexhaustible streams the most abundant and most marvelous Baseline Shift Bevel Blackletter F
Font Management Gridlines
liquor that has ever flowed to relieve the thirst of men. Through it, Inner Glow Italic Justification Ker
God will spread His word; a spring of pure truth shall flow from it; Old Style Font Outer Glow Points
like a new star it shall scatter the darkness of ignorance, and cause a
light hithertofore unknown to shine among men.
—Attributed to Johannes Gutenberg, inventor and pioneer of the modern
printing era (1398–1468)
Chapter Highlights
This chapter examines:
The origins of typography and the modern use of electronic type in multimedia designs
Styles and classifications for electronic typefaces in graphic design
Tools and techniques for managing the appearance of text
Tools and techniques for controlling character and line spacing, text placement, and alignment
Ideas for maximizing the readability of screen text in multimedia projects
AN INTRODUCTION TO TYPOGRAPHY
The element of text is one of the most important components of a multimedia
experience. Text is the visual representation of intellectual thought as expressed
through a human language system. Whether it’s in the form of texting, hyper-
text, tweeting, email, snail mail, or notes on a napkin at lunch, text plays a
big role in our lives. The Age of Enlightenment was predicated on humankind’s
abil- ity to share ideas and information through formally agreed upon
conventions of writing. Even in a multimedia age, we continue to rely on text
as the primary means of recording, receiving, and transferring human
knowledge and ideas. So what is typography? How do you make informed
choices about fonts, spac- ing, and other typesetting and layout options?
What makes Times New Roman recognizable and look different from Arial?
Let’s start at the beginning and get a few terms and ideas straight.
Multimedia Foundations. DOI: 10.1016/B978-0-240-81394-3.00009-4
© 2012 Elsevier Inc. All rights reserved.
246 SECTION 3 Static Media
FIGURE 9.1
The shape of a tree is formed by the thoughtful arrangement of stylized text. Here, text is
creatively used as an element of both form and content.
Ahmad Faizal Yahya/Shutterstock.com
FIGURE 9.2
A font family includes
all of the variant styles
associated with a
particular typeface
(bold, condensed,
light, italic, oblique,
etc.). The “regular”
style associated with
a typeface is its
default look or
appearance without
any stylistic
enhancements. Many
typefaces have only
a regular style.
FIGURE 9.3
This replica of the
Gutenberg press is
on permanent display
at the Gutenberg
Museum in Mainz,
Germany. Mainz is the
birthplace of Johannes
Gutenberg.
While the process of setting the type for a single page was quite time-
consuming, the actual printing of the page was very fast, especially when
compared to the old- school methods of manual reproduction by human scribes
or of block printing.
Printing technology has evolved considerably since Gutenberg’s day. Metal type
has not wholly been replaced by electronic type, but much typesetting is now
done digitally, and digital type is far easier to use and manipulate. With
digital type, you have the option of outputting text to a printer, computer
screen, or virtually any other type of electronic display. With tens of thousands
of typefaces available (at least) and so many easy options for customizing
text, you might think good typesetting would be a snap. But good
typographic design requires knowledge and careful planning. Good design
creates an aesthetically pleasing layout that communicates effectively to the
user (the reader, the viewer, or the person interacting with the multimedia).
Writing involves the crafting of words, while typography applies the complex
rules of design to the presentation of words on a printed page or digital
screen to not only present written ideas but also set a tone, connect to users,
establish credibility, and set expectations.
FIGURE 9.4
Legibility and
readability are related.
LEGIBILITY AND READABILITY
What letters and Legibility and readability are related terms. Legibility refers to a typeface’s
words are easiest to character- istics and can change depending on font size. The more legible a
distinguish? Which typeface, the easier it is at a glance to distinguish and identify letters, numbers,
ones are easiest to
read?
and symbols. A number of factors affect legibility, such as x-height and
Illustrator: Susan counter size (the size of the open space either fully or partially enclosed in a
A. Youngblood letter). In Figure 9.4 (left), compare the letterforms in Playbill, Arial, Rosewood
and the other typefaces shown for legibility.
Text CHAPTER 9 251
Readability refers to how easy text is to read in context, not as isolated letters,
and depends on a variety of factors, including typeface characteristics such as
italics, font size, style, letter spacing, line spacing, alignment, background,
capitalization choices, and contrast. Legibility is a factor, of course: if characters
are hard to identify, words are hard to read. Good type designs are usually
invisible, mean- ing that the reader’s attention is fixed on the words and the
meaning of the text, rather than on the stylistic features of individual characters.
When typesetting is sloppy and ill-conceived, the readability of text diminishes.
When users begin to consciously notice details of a typeface over content,
readability is diminished. When users have to strain to read—and when their
eyes fatigue because of type- setting choices—readability is diminished. In
Figure 9.4 (right), compare the line spacing, styles, and letterforms in Bodoni,
Times New Roman, and Script type- faces shown for readability.
FIGURE 9.5
Garamond is
considered a high-
contrast font because
of the stark difference
in the width of the
strokes used to form
a single character.
Arial Black, on the
other hand, has a
consistent width for
each stroke segment.
252 SECTION 3 Static Media
FIGURE 9.6
The characteristics,
illustrated here, not
only define each
typeface but also help
determine a typeface’s
suitability for various
uses.
Illustrator: Susan
A. Youngblood
FIGURE 9.7
Serifs are decorative
accents added to the
end of a stroke.
Serifs are small marks located on the ends of a main character stroke (see
Figure 9.7). Serif typefaces contain serifs, while sans-serif typefaces do not.
Which is better for print, and which is better for electronic display? Even a
few years ago, experts would have said that serif typefaces were better for
printed materials such as books because the serifs help with horizontal flow,
or help guide readers’ eyes across long lines of text. Some of this research on
print readability has been called into question; at the most, the difference
made by serifs is small. Other research suggests that sans-serif typefaces are
better for electronic displays. Why? Serifs are small, so lower-resolution
displays don’t have enough pixels to cleanly render the serifs. Sans-serif
typefaces are usu- ally better for computer monitors and video. As you make
typeface choices, consider specific font characteristics that improve legibility,
not just categories. Also consider what your users expect in a given context,
which is often serif typefaces for print and sans-serif for electronic media.
CATEGORIZING TYPEFACES
Typefaces are generally classified into two main groups depending on
whether or not they contain serifs, and those groups can be divided many ways.
The clas- sification system discussed here is only one way to name and group
typefaces.
254 SECTION 3 Static Media
FIGURE 9.8
Typeface development through history. The examples are typical of their categories, but some
were created later than the earliest dates listed for a given category.
Illustrator: Susan A. Youngblood
With so many variables, some typefaces are hard to classify. And different sys-
tems use some of the same words, such as Gothic, to mean different things (see
Figure 9.8).
Serif Typefaces
There are six main groups of serif typefaces: Blackletter, Humanist, Old Style,
Tran- sitional, Modern, and Slab Serif. Serif typefaces are the industry standard
for body copy that’s printed in books, magazines, and newspapers. Even if new
research calls into question readability differences on paper, readers are used to
serif typefaces in these contexts. Digital technology can have trouble displaying
serifs, so unless you’re using a serif typeface designed for the digital display, such
as Georgia, you’re usually better off relying on serif typefaces for large headings
or for print media.
press, used Blackletter type. Gutenberg used a two-column design with forty-two
lines of text per page. The typeface could be read because the font was big,
but the Gutenberg Bible was quite large and heavy when finished.
As printing moved into southern Europe, the more efficient and elegant
forms of Roman lettering greatly influenced the design of metal type.
Humanist (also known as Venetian) typefaces were the result. Like Blackletter,
they are dark on the page and have sloped crossbars in the lowercase e, but
despite their low contrast, they were considered an improvement over their
Gothic predecessor. Humanist letterforms have more rounded and gentler
strokes and are much easier to read in small print than Blackletter, despite
having relatively small x-heights. Many of the serifs on lowercase letters are
slanted, and serifs tend to be bracketed: the serifs connect to the main stroke
with a curve, providing a smooth and gentle transition from the thick to thin
areas of a font. Soon after, italic typefaces were developed; they allowed
printers to put more words on a line and are still used for emphasis. Old Style
(also known as Garalde) typefaces are similar to Humanist typefaces but are
distinguished by a somewhat more vertical stress and horizontal crossbars in
the lowercase e. Old Style typefaces include Garamond, Bembo, and Caslon
and remain popular.
TRANSITIONAL TYPEFACES
As the name implies, the design attributes of Transitional letterforms fall some-
where between those of Old Style and what’s to come: Modern-era type. Com-
pared to Old Style type, transitional typefaces, such as Times New Roman and
Baskerville, have higher contrast, a larger x-height, vertical stress, wider brack-
eted serifs, and generally wider letterforms.
Times New Roman is one of the most ubiquitous and well-known typefaces
in this category. After being criticized for sloppy printing and the lack of
readabil- ity in their newsprint, the London-based newspaper The Times
commissioned the Monotype Corporation in 1931 to develop a new typeface.
Released a year later, Times New Roman quickly evolved into the industry
standard typeface for the newspaper, magazine, and book publishing
industries. Today, Times New Roman is still widely popular and is one of a
handful of universal fonts that’s included with nearly every computer and
operating system in the world.
MODERN TYPEFACES
Modern (also known as Didone) typefaces represented the first noticeable depar-
ture from typography’s historical dependency on pen-based letterforms.
Their designs have an extremely high contrast, a small x-height, thin horizontal
serifs with little to no bracketing, and vertical stress in the rounded strokes.
Modern fonts are not very readable when reduced to small, lengthy sections of
text. Their contemporary features make them a better choice for title text and
headers.
SLAB-SERIF TYPEFACES
Up until the time of the Industrial Revolution, typefaces were largely
256 SECTION 3 Static Media
designed for setting small body type in newsprint and books. With
industrialization
Text CHAPTER 9 257
came the increased need to promote and advertise goods and services for public
consumption using large letter type. Slab-serif (also known as Egyptian)
typefaces were designed specifically for this type of application. Bold and
eye-catching, slab-serif typefaces are a good choice for posters, flyers,
billboards, and other large format media that demand immediate attention and
text recognition. Slab- serif typefaces have low-contrast strokes, thick horizontal
serifs that can be either squared-off or slightly bracketed, and vertical stress in
the rounded strokes.
Sans-Serif Typefaces
Sans is a French word meaning “without,” thus sans-serif (also known as lineal)
typefaces are those that are literally without serifs; they are sometimes also
referred to as gothic or grotesque, because some earlier typeface designers found
them unattractive. Sans-serif type was used to emboss pages for blind readers
beginning in 1789 and was developed for printing for sighted readers in
1816. The popularity of sans-serif typefaces boomed in the 20th century. Type
design- ers strove to create basic letterforms, devoid of the excessive design
elements of modern and slab-serif type. Sans-serif typefaces obviously have
no serifs, and usually have uniform strokes with little to no contrast and a
vertical stress in rounded strokes. These typefaces are also broken into
categories: Humanist for varieties that have an angled stress, nongeometric
counters, and contrast; Geometric for varieties that have vertical stresses,
geometric counters, and no contrast; and Transitional for those in between.
FIGURE 9.9
These decorative fonts
were downloaded
under a freeware
license
from urbanfonts.com.
UrbanFonts is a
popular font-sharing
website that provides
access
to thousands of fonts.
Through this site, and
many others like it,
users can download
fonts for free (as
freeware) or on a trial
basis (as shareware).
In the case of
shareware, the cost is
often much lower than
commercial rates.
While known mostly for
its large collection of
free fonts,
258 SECTION 3 Static Media
urbanfonts.com also
sells fonts.
Text CHAPTER 9 259
Sans-serif typefaces are ideal for text headings and titles in print and for use
in electronic media where lower resolution can make serifs hard to render
cleanly. Because of this, sans-serif fonts are often the best choice when
designing body copy for the Web or other channels of electronic distribution.
Decorative Typefaces
Decorative typefaces connote a sense of mood, emotion, or attitude (see Figure
9.9). They have personality, which is great for attracting attention but does
little for enhancing the readability of the text in small form. Decorative typefaces
are charac- terized by unusual features intended to add splash and pizzazz to a
design. Because they’re purposely designed to draw lots of attention to
themselves, they should be used in moderation.
Script Typefaces
Script typefaces are among the easiest letterforms to categorize because they so
clearly attempt to emulate the cursive style of handwriting, or the artistic appear-
ance of calligraphy. Script typefaces are slanted, and a stroke connects adjoining
letters and preserves linear flow. Script typefaces are popular with
commercial printers when they need to create a formal text design such as in
a wedding invitation or certificate of achievement, when readability is less of
a concern than the overall feel of the document. Because of their thin strokes
and angled form, these typefaces generally have a weak appearance when
rendered out to electronic displays.
FIGURE 9.10
The Webdings font includes this fun assortment of symbol glyphs. Many programs, including
Microsoft and Mac operating systems, include a character map that allows you to view, select,
and insert special characters and symbols into an open document. The Glyphs window shown
here (top) is available from within Adobe Illustrator.
In order for a web browser to accurately display text in a web page, all of the
fonts used in its design must reside on the client computer. When the browser
is unable to locate a font that’s specified in the HTML code, it will perform a
font substitution, swapping out the intended typeface with a local one. Font
substitution can significantly change the look of a page from what was originally
intended. To ensure font compatibility with other systems, select a font from
one of the following families: Verdana, Georgia, Arial, Courier New, Times
New Roman, Trebuchet MS, Comic Sans, and Symbol. These are standard
fonts that come with all Windows and Mac computer systems. They’re also
considered to be Web-safe, meaning they are supported by all of the most
popular web brows- ers (see Figure 9.11).
Relying solely on safe fonts can limit your creativity. As an alternative, text can
be saved as a graphic image and uploaded to the Web in JPEG or GIF format.
With the typeface embedded inside the text graphic, users don’t need to have
the matching font installed on their local computers to view it. However, use
images of text only for titles, headers, and display text: search engines do not
have the ability locate text when it’s embedded in a graphic, and users with
disabilities need you to provide alternative text for every image, which can be
cumbersome if you have a text-heavy image. If you choose to create limited
text in images, consider using an OpenType font: these fonts are scalable and
crossplatform, so they’ll look right if you move between a Mac and PC during
your project.
FIGURE 9.11
Like most web authoring programs, Adobe Dreamweaver CS5 displays a relatively small list of
fonts to choose from. The fonts shown are ones considered by Adobe to be “Web-safe.”
262 SECTION 3 Static Media
Tech Talk
Category Characteristics
A typeface’s characteristics help determine which category it fits in and what it is best used for.
Humanist Sans
Serif no serifs; angled stress print headings and Optima Gill
digital text; Verdana SansCalibri
was designed for Verdana
digital use
Transitional Sans
no serifs; low contrast; print headings and Helvetica Arial
Serif
almost vertical stress digital text
Geometric no serifs; based on
advertising Futura Century
geometric forms;
Gothic
vertical stress
Italic angled, but the roman
emphasis, Didot Italic Arial
form has been adapted,
calligraphy Italic
not just slanted
Decorative letterforms have deco-
limited eye- Mesquite Std
rative touches
catching text Curlz MT
Princetown LET
Script a subset of Decorative
formal Edwardian Script
modeled after hand-
invitations ITC Lucida Hand-
writing strokes
writing
Symbol symbols often not
for special charac- Symbol
available on the key-
ters (e.g. - language, Wingdings Zapf
board
currency, mathe- Dingbats
matical, geomet- ric,
and phonetic
symbols)
GREAT IDEAS
Adding to Your Typeface Collection
Typefaces can be added to a computer system (as fonts) in a number of different ways. An assortment of basic fonts is inc
264 SECTION 3 Static Media
FIGURE 9.12
Linotype is a
commercial type
foundry that
specializes in the
design of commercial
typefaces for
professional designers.
Founded in 1886,
Linotype designers
have created a large
collection of fonts
including familiar
ones such as Gil Sans,
Helvetica, Optima,
Palatino, and Papyrus.
Text CHAPTER 9 265
Adding emphasis to key words, headings, or phrases lets the user know that
there is a sense of order and importance in the way that information is pre-
sented. When all of the text in a page looks exactly the same, the design
becomes visually flat and viewers are left to wander aimlessly through a sea
of monotonous pixels. On the other hand, if you emphasize too many text
ele- ments at once, the design may appear visually cluttered and confusing.
In other words, if you emphasize just about everything, you emphasize nothing.
As with all things in design, moderation and balance are important. Text
emphasis is most often accomplished by varying the style, size, and color of
a typeface (see Figure 9.13).
FIGURE 9.13
The Character panel in Adobe Photoshop CS5 can be used for modifying the appearance of
selected text in the design window Source: Adobe Photoshop CS 5.
Source: Adobe Photoshop CS 5
266 SECTION 3 Static Media
FIGURE 9.14
The difference between true bold and italic styles and faux effects is quite apparent when you
compare them side-by-side.
from what the designers of the font intended (see Figure 9.14). Professional
designers prefer the authentic look of true font styles, because the size, shape,
and spacing of each letterform are custom designed to be consistent with the
overall look of the typeface; they tend to avoid faux styles because they are
aesthetically less pleasing than the real thing. Some typefaces have more true
font styles than others.
FIGURE 9.15
Type can be condensed, expanded, scaled, and stretched a number of different ways to achieve
a desired effect.
of type, adjust the percentages up or down in the appropriate entry field. Verti-
cal and horizontal scaling are virtual effects that merely stretch out type. As a
result, they are not nearly as pleasing to the eye as a true condensed or expanded
typeface.
Font Size
Digital fonts, whether used to create materials for print or for the screen, are
measured in real physical units that correspond to their actual size when the text
is printed on paper. But when fonts are displayed in an electronic medium,
Text CHAPTER 9 269
the
270 SECTION 3 Static Media
actual size of text will vary according to the resolution (in pixels) of the monitor
used to view it. You can set font size in several ways, but the point system is the
most common method for measuring type. For print, it’s customary to set the
font size to 10–12 points for body copy. Research and experience have
shown these to be good point sizes for text documents when read by most
people with normal vision. But just how big is 12-point font?
To begin with, points are a unit of measurement used to define the vertical
height of a typeface. Originally, point size referred to the vertical height of
the metal block with the raised letterform. Regardless of a character’s shape and
size, the block for each character in a typeface had the same height. An
uppercase Z, though, naturally consumes more surface area on its metal block
than does a lowercase i. And for both metal and digital type, point size
reflects the distance from the top of the tallest ascender to the bottom of the
lowest descender, plus a bit of additional space on either end. There are 12
points in a pica, and 6 picas in an inch. So there are 72 points in an inch. It’s
worth mentioning that many monitors, especially Macs, display 72 ppi
(points per inch). Does this mean that using a 72-point font will always lead
to the production of letters that are exactly one-inch tall? Well no, not really. As
long as fonts are limited to display purposes only on electronic screens (as
opposed to printed out), the fact that points can be converted to inches really
doesn’t matter. For the multimedia producer, it’s much easier to think of font
size in relative rather than absolute terms. The readability of an impression
made by different fonts set in the same point size can vary significantly in
terms of their visual weight and impact. Even though different fonts of the
same point size may use virtual blocks with the same height, one’s capital
letters may be shorter within the block. Or another’s x-height may be greater,
making it seem larger. Of course, some fonts, such as Georgia, may have
wider letterforms too (see Figure 9.16).
FIGURE 9.16
All five lines of type
pictured here were set
to a 12-point font size
in Adobe Illustrator
CS5. However, the
printed length of each
line varies because of
variables associated
with the design of the
each typeface being
used. The display size
of the fonts shown
here was enlarged for
comparison.
Text CHAPTER 9 271
When creating text in graphics software, always be sure to save it to a layer that
is editable so you can resize or revise it later. In graphics software, type is
initially created as a vector object, allowing it to be scaled to virtually any
size without losing detail. Text can be resized with precision by entering or
selecting a discrete point size. When a graphic is saved in a bitmap file format
such as JPEG, the text elements within it are permanently rastorized, or
changed to pixels rather than lines, and can no longer be edited with the type
tools or resized without result- ing in a loss of image quality and detail.
Font Color
When selecting a font color, be sure to pick a contrasting shade that stands
out from the background. Combining a dark-colored font with a dark
background, or a light-colored font with a light background, leads to poor
contrast, and the text will be difficult to read (see Figure 9.17). Dark type on a
light background is gen- erally easier for people to read than light-colored type
on a dark background. Even electronic books are normally created with black
type on a white background.
When you do choose splashes of color, be careful about the combinations.
Some color combinations, such as green and red, seem to vibrate, so they are
hard to look at and strain users’ eyes. Also, some of your users may be color-
blind or have color deficiencies. There are several kinds of problems that affect
seeing color, and some color combinations are poor choices. For instance, a
person may perceive all red as black, so that the black-on-red website mar-
quis you planned won’t work (unless you want the user to see a big black
FIGURE 9.17
Color contrast affects
readability.
Illustrator: Susan
A. Youngblood
272 SECTION 3 Static Media
box). Whether you’re working on graphical text or web text, check your
color choices with a tool such as Vischeck.com’s online colorblindness
simulator.
Baseline Shift
When composing type in a word processing or graphic design program, the soft-
ware automatically conforms all of the text in a row to a common baseline.
Much of the time this works just fine. However, there are other times when
the visual aesthetic can be best served by shifting the base of certain
characters to a new position. This is particularly true when you attempt to
combine different font families, styles, or sizes. You can shift the baseline to
move text one point at a time above or below the existing baseline (see Figure
9.18).
Superscript/Subscript
In order to accurately reflect the shorthand notation for special formulas and
character symbols, it’s sometimes necessary to apply subscript or superscript
formatting to text. Superscript characters have a smaller point size than the
sur- rounding text and are shifted upward above the baseline. Ordinal
numbers are often displayed with a superscript modifier (1 st, 2nd, 3rd, etc.).
Other uses for superscript formatting include footnote references 1, compass
bearings (20o W), and mathematical exponents such as 10(2) and (an)1.
Subscript characters also have a smaller point size, but are shifted downward
below the baseline as in H2O, the chemical formula for water.
Anti-Aliasing
Because screen fonts are rendered out on screen using square pixels, the edges of
the diagonal and curved strokes in a typeface often appear jagged. This stair-step
effect is called aliasing and is an undesirable consequence of digital typography.
Anti-aliasing is a technique used in raster (bitmap) editing to smooth out the
edges of type. When anti-aliasing is turned on, transition pixels of an intermedi-
ate color are added between the jagged edges of a stroke and the surrounding
area. The result is a distinctly smoother edge. Anti-aliasing works best on
large type and is not generally recommended for font sizes smaller than 10
points because it reduces the readability of text.
FIGURE 9.18
Typically, all type in a row of text sits uniformly on a common baseline (top). Shifting or varying
the baseline (bottom) is a creative choice by the designer for achieving a particular visual effect.
1
A superscript footnote reference like the one inserted in the text above, corresponds to a citation or
Text CHAPTER 9 273
note (like this one) placed at the bottom of the page or at the end of a chapter.
274 SECTION 3 Static Media
Tech Talk
FIGURE 9.19
Suitcase Fusion 3 is a crossplatform font management program compatible with both Mac and Windows.
Text CHAPTER 9 275
FIGURE 9.20
Font Book is Apple’s proprietary font manager utility for computers running OS
X. The Baskerville font family is highlighted here as one of 259 resident fonts in the fonts collection.
Leading
Leading (pronounced ledding) is used to define the amount of space between
verti- cally adjacent lines of text (see Figure 9.21). The term originated during
the days of manual typesetting when thin strips of lead of various widths were
used to separate the rows of metal typeface. In digital typesetting, leading is
measured in points as the distance between the baseline of one row and the
baseline of the next. When lines are spaced too closely together or too far apart,
readers have trouble advanc- ing their eyes to the next line, and readability
diminishes. The computer-generated spacing that’s assigned to type by the
graphics program is not always appropriate and should never be blindly
accepted. For example, large x-height fonts—such as most sans-serifs—usually
require additional leading. Extra leading is also needed for longer lines of text
and for paragraphs set in boldface type.
276 SECTION 3 Static Media
FIGURE 9.21
Which of these
blocks of text is
easier to read? Which
is the most difficult
to read? Why?
paragraphs; it disrupts reading and reduces white space within the page
layout. Right-aligningtext with a left-ragged edge can add variety and interest to
a design. However, since readability is diminished with this technique, its best to
limit its use to short paragraphs or sentences. Centered text produces a
symmetrical look that’s visually balanced but difficult to read, since both the
left and right edges are ragged. Center alignment is not recommended for body
copy and only spar- ingly with headlines and other short sections of text (see
Figure 9.22).
It’s never a good idea to rely upon your eye or your skill with a mouse when
aligning and distributing objects—evenly spacing them out vertically or hori-
zontally—within a layout. Alignment needs to be precise, otherwise you may
end up introducing irregularities into the design that are visually distracting
to the viewer. You can also align precisely by using gridlines and snapping
tools (see Figure 9.23). A grid is a matrix of evenly spaced vertical and
horizontal lines that are superimposed overtop of the design window. Most
multimedia design and editing software includes gridlines or snapping to
make it easier to keep objects and text perfectly aligned. If you use snapping, as
you move a layer close to a gridline, the layer snaps into place.
FIGURE 9.22
(1) Ragged-right or left justified (most common for body copy); (2) ragged-left or right justified;
(3) centered text; and (4) justified left and right.
FIGURE 9.23
Most design programs have a set of alignment buttons or menu commands for aligning
selected text and objects within the design space. The icons shown here will be familiar to users
of Adobe Photoshop CS5.
278 SECTION 3 Static Media
FONT TRANSFORMATIONS
You can transform type into an element of design to grab the attention of the
viewer or to add interest and variety to a page layout (see Figure 9.24).
Visual effects can change the appearance of the shape, fill area, and stroke of a
typeface in an endless variety of fun and interesting ways (see Figures 9.25 and
9.26). Be careful, though: if the effects aren’t subtle, they can be distracting.
And many effects, such as drop shadows and beveling, can reduce text’s
readability.
The following visual effects are fairly common and can be easily applied to text
by a designer using a program such as Adobe Photoshop or Illustrator.
■ Color gradients and pattern fills. While the conventional type in body copy
is usually filled with a single color, type can also be filled with a gradient
of colors, a pattern texture, or an image.
■ Warping. Warping bends and distorts a typeface. Text can be arched,
inflated, squeezed, twisted, and manipulated a number of other ways to create
a vari- ety of text-based shapes.
■ Drop shadows. Adding a drop shadow is one of the easiest things you can
do to add interest and depth to text objects and backgrounds. When
applied to text, a drop shadow gives the illusion that words are floating above
the back- ground. The more pronounced the drop shadow is, the greater
the distance will appear between the text and the background. The perceived
foreground- background position of objects in a composite can be
manipulated in part by applying different drop shadow settings for each
object.
FIGURE 9.24
An assortment of some of the most common effects you can apply to text for enhancing visual contrast and impact.
First Column: warped text, bevel and emboss; Second Column: reversed type, stroke/outline, drop shadow, and inner
Text CHAPTER 9 279
glow; Third Column: pattern fill (grass), gradient fill (three tone color), image fill (flag), and outer glow.
280 SECTION 3 Static Media
FIGURE 9.25
Top-left: Plain white text (no font transformations applied). Top-right: A drop shadow is applied. Bottom left: A stroke/
outline is applied. Bottom-right: A two-color gradient fill and drop shadow are applied. Which version do you think has
the best contrast and visual appeal?
FIGURE 9.26
A clipping mask was used in Adobe Photoshop CS5 to create this colorful font-fill effect.
Text CHAPTER 9 281
■ Stroke. A stroke is a colored outline that’s placed around the exposed edges
of type. Adding a stoke effect can greatly improve the contrast between text
elements and the background. An inside stroke cuts into the fill area of type,
while an outside stroke expands outward toward the background. A center
stroke expands it in both directions at once. A stoke can be a solid color,
a gradient color, or a pattern texture.
■ Bevel and emboss. Beveling rounds off the edges of type, giving it the
sculpted, graceful look of a raised letterhead, making it appear three-
dimensional. Embossing produces the opposite effect, making type appear
stamped or pressed into the background. The illusion created with these
effects varies greatly depending on the settings and the color of background.
■ Inner and outer glow. An inner glow applies colored shading to the inside
edge of type to add internal depth or highlights to the fill area. Outer glow
creates a halo effect around the outside of type to improve contrast and back-
ground separation.
Layer styles can be mixed and matched in what seems like an endless number of
combinations. The potential for creativity is virtually infinite.
Tech Talk
Less Is More
There’s a popular adage in design circles that says, “less is more.” With such
an abundance of fonts, styles, and visual effects at your disposal, you may be
tempted to throw everything in the pot, wave a magic wand, and see what comes
out. Less is more suggests that the best styles and effects are often those that are
subtle or moderated. You shouldn’t have to SHOUT to get people’s attention. At
the right moment, a whisper can be even more effective. So as a final cautionary
note, avoid going overboard with all the tools you now have at your disposal.
The less is more principle implies that a simple and elegant design that commu-
nicates effectively is much more valuable than a visual-effects showpiece with
no apparent focus or message.
277
There are always two people in every picture: the photographer and the Key Terms
viewer. 35mm
Angle of View Aperture
—Ansel Adams, Photographer (1902–1984) Aperture and Shutter Priority Mod
Auto and Manual Focus
Camera RAW Consumer, Prosumer
Chapter Highlights Decisive Moment Depth of Field D
EXIF
This chapter examines:
Exposure Exposure Modes F-Stop
How digital cameras are classified according to their operational features and intended use
Film
The purpose and function of the imaging chain and each of its basic components Film Speed Fisheye Lens Flash
The variables affecting the proper exposure of a digital image Focal Length Focal Length Multipl
The use of fully automatic, semi-automatic, and manual shooting modes Image Resolution Image Sensor Im
Strategies for organizing and managing digital image files Iris ISO
Landscape Mode Lens
Macro Lens Memory Card Metadat
PHOTOGRAPHY
Photography is the process of fixing an image in time through the action of light.
In traditional chemical processing, photographic images are created by exposing
a light-sensitive emulsion on the surface of film to light in a controlled environ-
ment. While some people still shoot with film, the vast majority of the world has
crossed over to digital photography, an electronic medium that renders pictures
using a digital image sensor. The demise of film was highlighted in 2005 when
the Kodak Corporation reported that digital sales accounted for 54% of annual
revenue, exceeding film sales for the first time in company history. Sadly, this
legendary company founded in 1880, and that pioneered the development of
photographic film, filed for Bankruptcy protection in January 2012.
Digital photography offers instantaneous results, producing image files that
are easily transferable, and adaptable, for a wide range of multimedia
products (see Figure 10.1). The rapid adoption of digital cameras over the
last decade can be attributed to many factors, including improvements in
reliability and ease of use; better image resolution and print quality; higher-
density storage solutions;
and Zoom
Normal Lens Optical/Digital greater compatibility
Photography withMode
Portrait computer
Primaryhardware and software
Lens Reciprocity systems. Digital
Law Red-Eye
Selective Focus Shutter cameras
Sports/Action
come Mode Stich Mode
in three mainTelephoto
varieties:Lens Tripod prosumer, and professional.
consumer,
TTL
While the dividing line between each level is somewhat arbitrary, manufacturers
White Balance Wide-Angle Lens Zoom Lens
use the terms to distinguish between low-, medium-, and high-quality equip-
ment within a product line.
Consumer Cameras
At the lowest end of the consumer spectrum are point-and-shoot cameras
designed for users with little to no background in photography. The point-
and- shoot camera automatically determines the appropriate focus and
exposure set- tings. Most of these cameras have a built-in flash and are
largely menu-driven, with very few controls located on the camera body
itself (see Figure 10.2).
Some consumers prefer the body style, features, and quality benefits of a
digi- tal single lens reflex camera (DSLR) over the scaled back point-and-
shoot. In a DSLR, a mirror located behind the lens directs the image to an
optical view- finder, allowing the photographer to compose and focus a scene
while view- ing it through the lens (TTL). When the shutter is released, the
mirror flips out of the way, allowing the image to strike the image sensor.
Experienced photographers prefer to compose through-the-lens because it’s
more precise. Consumer-grade DSLR cameras normally have a permanently
mounted lens, which, while usually much better than the lens on a point-and-
shoot camera, may not be as good as that of a prosumer or professional
lens. Regardless of the style, consumer cameras are designed for mass-
market appeal and are engineered for size, portability, ease of use, and cost.
They are not typically designed to give the user easy access to camera
settings. The quality of these
FIGURE 10.1
Photography is continuously evolving, both as an art and a science. While the traditional ways often seem archaic, they
continue to provide high-quality images. Likewise, while contemporary ways may lack marginally in quality, they make
up for it in alacrity and ease of postproduction. Pictured here (L-R), Molly Bartles, a photojournalist for the Courier & Press
in Evansville, Indiana, shoots with a professional-grade DSLR, as Chris Walker, a documentary photographer who teaches
at the University of Southern Indiana, is shown with his 80-year-old wooden 8 10. Each is a specific tool that, even in
contemporary America, has its place. Photographers: Chris Walker (left) and Josh Reuck (right).
Photography CHAPTER 10 279
FIGURE 10.2
The Apple Store website claims that the iPhone 4S “just might be the only camera you’ll ever
need.” Is the market for the low-end point-and-shoot camera disappearing forever? While
resolution is just one benchmark of picture quality, the fact that the iPhone comes with an
8-megapixel image sensor is impressive. Why carry around a camera and a phone when
one device may do the trick?
Source: apple.com/iphone
cameras has improved tremendously over the last decade. While one of the
differences between consumer and prosumer digital cameras used to be the
number of megapixels (the number of light-sensitive elements on the sensor,
measured in millions), there is no longer as much difference, and it is not
uncommon to find 10- and even 14-megapixel consumer-grade cameras. A high-
resolution image from a 10-megapixel-camera will easily print up to an 11
14-inch image.
Tech Talk
Video Cameras
While video technology will be discussed in more detail later on, it’s worth
mentioning here that digital video cameras and digital still cameras share
simi- larities in basic operation, including controls, shooting modes, and
exposure settings. Because of this, many of the concepts in this chapter can
be applied to the use and operation of video cameras. Given their innate
similarities, it should come as no surprise that video cameras today often
allow users to capture still images, while many still cameras allow
photographers to shoot moving images and sound. As a general rule,
however, it’s best to use a camera for what it was designed for. Since a video
camera is optimized for moving images and sound, its still image processing
capabilities and resolution options are generally not as good as a high-end
digital camera, though high-end video cameras such as the RED are an
exception. Similarly, most still cameras are not ideal for captur- ing video,
particularly from an audio perspective. As with many things, the best results
come by choosing the right piece of equipment for the right job (see Figure
10.3).
Tech Talk
FIGURE 10.3
Standard camera formats include DSLRs (approximately 16 24mm up to medium format), SLRs (24 36mm), medium
format (2¼" square), and large format (4 5 inches and above). In this image, the lens used to make the original 4 5
version was a 90mm, which is wide angle for that format. But as the highlighted boxes illustrate, that same lens would
render a “normal” angle of view for 2¼", and a slight and medium telephoto for 35mm and a DSLR with an FLM, respectively.
The equivalent lens needed to duplicate this image’s angle of view with a prosumer DSLR would be approximately 20mm.
Photographer: Chris Walker.
The Lens
In still cameras and video camcorders, the lens is an integrated optical system
made up of one or more individual elements of ground glass or molded
plastic. The lens is mounted on the front of the camera and is designed to
capture and
Photography CHAPTER 10 283
FIGURE 10.4
A DSLR camera with an assortment of interchangeable lenses.
manipulate light reflected from objects in the camera’s line of sight. The lens
controls the magnification of the subject being photographed and affects the
sharpness and clarity of the image (see Figure 10.4).
PRIME LENSES
Lenses with a single focal length are called primary or fixed focal length lenses.
With a fixed focal-length lens, the only way to affect angle of view is to
physi- cally change the distance between the camera and the subject. Serious
photog- raphers often own an assortment of prime lenses with a variety of
focal lengths, enabling them to manipulate angle of view by simply swapping
out the lens. Prime lenses are often classified into one of six main categories:
wide-angle, tele- photo, normal, novelty (macro and fisheye), and super-
telephoto.
Tech Talk
■ Wide-angle lenses have a relatively short focal length (18–35mm for full
frame, 12–24mm for APS-C) resulting in the wide angle of view. Wide-angle
lenses, or short lenses as they’re sometimes called, are often used for
shooting landscape panoramas and vistas where the primary emphasis is on
establishing a wide overview of the scene (see Figure 10.5).
■ Telephoto lenses have a long focal length (90–300mm for full frame, 60–
200mm for APS-C) resulting in a very narrow angle of view. Telephoto,
or long lenses, can magnify distant objects, making them appear much closer
FIGURE 10.5
Wide-angle lenses are often used to capture broad vistas, yielding images such as this one.
Photographer: Chris Walker.
FIGURE 10.6
To understand the importance of lens selection, find a subject, like the first red door on this church, and photograph
it up close with a wide angle lens. Then measure the distance from your position to the primary subject and double
that distance. From that farther spot, zoom in to twice the focal length of your initial image and retake the picture.
Your primary subject will retain its original size, but your secondary subjects will become more prominent. Good
photographers find their ideal position, and then select the lens that will best depict the scene from that location.
Photographer: Chris Walker.
Photography CHAPTER 10 285
than they really are. These are the favorite lens of bird watchers, sports
pho- tographers, and naturalists for obtaining close-up shots of a subject
from a distance (see Figure 10.7-Left).
■ Normal lenses have a medium focal length (40–60mm for full frame, 28–
40mm for APS-C) resulting in an angle of view that falls somewhere in
between the extremes of wide-angle and telephoto. Normal lenses are a
safe bet for general-purpose shooting activities that do not require extreme
close- ups or wide-shots of the subject matter.
■ Novelty lenses have descriptive names like macro and fisheye, and feature
unique optical characteristics designed for nontypical shooting situations.
■ Macro Lenses can acquire focus when positioned only a few inches away
from the subject. Macro lenses produce high-quality images with low image
dis- tortion and are useful for shooting extreme close-ups of small objects like
insects or flower blossoms (see Figure 10.7-Right).
■ Fisheye lenses have an extremely short focal length (8–15mm for full frame,
11mm or less for APS-C), offering an angle of view as high as 180º.
However,
FIGURE 10.7
Because of their magnification, telephotos and macros are more likely than other lenses to add camera shake to your
image. To help avoid this, a good starting point is to convert the lens’s focal length into a fraction and use that as a
shutter speed—a 200mm lens would become 1/200th, so it should be hand-held at 1/250th or faster. Photographer: Chris
Walker.
286 SECTION 3 Static Media
Tech Talk
ZOOM LENSES
A zoom, or variable focal-length lens, can be adjusted to any focal length within
a set optical range. For example, using a zoom lens with a range of 28–
300mm, the focal length can be set to wide angle (28mm), telephoto
(300mm), or any- where in between. Zoom lenses allow the photographer to
quickly change the angle of view without having to swap out the lens. With
this type of lens, the photographer zooms in to acquire a narrow angle of
view and zooms out to compose a wide-shot. The zoom ratio indicates the
magnification capability of a variable focal-length lens. For example, a 3 or
3:1 zoom ratio means that the size of the image at the longest focal length
setting (telephoto) is three times greater than the size of the image at the
shortest setting (wide angle). High- quality zoom lenses are very expensive.
While their optical quality is generally inferior to that of primary lenses, the
speed and convenience of zooming can be a worthwhile tradeoff. Zoom
lenses are included on most point-and-shoot still cameras as well as on all
consumer, prosumer, and professional video cameras.
The Iris
The iris is an adjustable plastic or metal diaphragm that regulates the amount
of light striking the image sensor (see Figure 10.8). The iris is typically located
between the optical elements of the lens. When shooting images that are dimly
lit, the iris can be opened up to allow more light to pass through to the image
sensor. For brightly lit subjects, the iris can be constricted or “stopped down” to
reduce the intensity of the light and keep the image from being overexposed. The
term aperture is sometimes confused with the iris, but it refers to the actual hole
or opening created by the iris and not to the device itself; so when people talk
about adjusting the aperture, they are usually talking about adjusting the iris to
change the size of the aperture.
Photography CHAPTER 10 287
FIGURE 10.8
An iris is typically
comprised of a series
of metal blades that
fold together to form
a circle, inside your
lens, through which
the image passes.
Photographer: Chris
Walker.
Table 10.2The Diameter of the Aperture is Obtained by Dividing the Focal Length of the Lens by the f-n
F-STOP
On professional lenses, the aperture is adjusted by turning the innermost ring
on the outside of the lens housing. A series of f-stop numbers are printed on
the outside surface of the ring indicating the size of the aperture. While shown
as a whole number, an f-stop unit is actually a fraction used for calculating the
physical diameter of the aperture. For example, an f-stop setting of f/16 means
that the diameter of the aperture is equal to 1/16th the focal length of the lens
(see Table 10.2). To calculate the aperture of a 50mm lens set to f/16, you would
use the following formula:
Typical f-stop positions include: f/32, f/22, f/16, f/11, f/8, f/5.6, f/4, f/2.8, f/2,
and f/1.4, though you typically won’t find a single lens that covers the entire
range. Because the f-stop number is a fraction, the size of the aperture
actually decreases as the f-stop number increases.
Opening the aperture by one full f-stop (e.g., changing from f/11 to f/8), doubles
the size of the aperture and the amount of light striking the image sensor.
Like- wise, closing the aperture by one full stop (e.g., changing from f/1.4 to
f/2), reduces the size of the aperture and the amount of incoming light by half.
Some lenses allow the aperture to be adjusted in half-stop or third-stop
increments for even greater control of light exposure.
The Shutter
The shutter is a movable curtain, plate, or other device that controls the
amount of time that the image sensor is exposed to light. When the
shutter button is depressed, the shutter opens for an instant of time allowing
light to strike the surface of the image sensor. The amount of time that the
shutter remains open is referred to as shutter speed and can commonly range
from 1/8000th of a second to several full seconds. Fast shutter speeds in the
neighborhood of 1/250th or 1/500th of a second are good for freezing action
in a well-lit scene (see Figure 10.9). Much slower shutter speeds are required
when shooting under low-light situations.
FIGURE 10.9
The exposure time
for this image
was
approximately
1/1000th of a second.
Faster shutter speeds
are available on all
DSLRs, but it’s fairly
rare that we use them.
Photographer: Chris
Walker.
Photography CHAPTER 10 289
GREAT IDEAS
The Image Sensor
A digital camera’s image sensor is a small electronic chip used to register the intensity and color of ligh
FILM SPEED
One of the challenges of digital photography is that much of the terminology is
rooted in film. One of the best examples of this is the continued use of the ISO
Film Speed system. Originally based on a film’s ability to respond to light, the
ISO rating is now used to describe a sensor’s ability to respond to light. The ISO
designation follows a logarithmic scale, which means that each jump to a higher
ISO number results in a doubling of the film’s light-sensitivity (see Table 10.3).
This corresponds nicely to the f-stop scale, which works in precisely the same
manner. All things being equal, increasing film speed by one ISO level has the
same effect on exposure as opening the aperture by one full stop. As film speed
increases, however, the sharp- ness and clarity of an image decreases, and your
image will have a grainier, or “nois- ier,” appearance (see Figure 10.10). For the
cleanest images, it’s best to shoot with the lowest possible film speed setting.
Unless you choose to manually override the ISO setting of the camera, your
camera will usually set your film speed automatically based on lighting
conditions, flash settings, and other selected exposure settings.
FILM SIZE
With photographic film, the surface area of a captured image has a fixed size,
based on the type of camera being used (35mm, etc.). In digital cameras, the size
of the image sensor is defined by the number of light-sensitive pixels it contains
and the
FIGURE 10.10
Electronic and analog media share many characteristics, including the degradation of image quality at higher ISOs.
As the enlarged view illustrates, the discrepancies between ISOs 200 and 3200 are quite apparent. As technological
improvements continue to be made, the distance in quality between low and high ISOs continues to shorten.
Photographer: Chris Walker.
physical dimensions of the chip. Generally speaking, the larger the chip, the
larger the acquired image based on its pixel dimensions. For example, a 6-
megapixel camera has roughly 6 million pixels and is capable of rendering
images up to a maximum resolution of 2,816p 2,112p. A 3.2 megapixel
camera has roughly half as many pixels and a maximum resolution of 2048p
1536p. While each camera’s image sensor is constrained to a fixed maximum
size, the resolution settings of the camera can be changed to produce a lower
resolution image.
FILE SIZE
Image file size is affected by the resolution setting of the camera and the file for-
mat used when saving a picture to memory. Consumer cameras typically process
and format images using the popular JPEG compression codec. JPEG compres-
sion reduces the size of image files as well as the time it takes for the camera
to process sensory data once the picture is snapped. Prosumer and
professional cameras also support JPEG encoding, but both usually also
provide the option of saving images in uncompressed formats like TIFF or
Camera RAW. While these formats produce files that are significantly larger,
they offer the advantage of preserving most or all of the original sensor data.
The Camera RAW format records a completely unprocessed version of the
image as obtained by the sensor at the point of exposure. Professional
292 SECTION 3 Static Media
photographers
Photography CHAPTER 10 293
prefer the RAW file format because it allows them to retain maximum
control of image processing during editing. JPEG and TIFF formatted files are
processed prior to encoding, limiting the type of adjustments that can be
made in image editing. Camera RAW is a proprietary standard that requires
special software or plug-ins from the camera manufacturer in order to edit the
image. While Cam- era RAW is a technically superior format, TIFF and JPEG
images are much easier to work with, and their quality is acceptable for the
vast majority of multimedia applications.
While the number of options varies by model and manufacturer, most digital
cameras allow you to specify the amount of compression you want to apply
to a JPEG image. For example, Canon offers three choices called normal, fine,
and superfine. As the name implies, superfine produces the highest quality image
by applying the smallest amount of compression. There’s a dramatic
difference between the file size of an image acquired at the smallest
resolution and low- est quality JPEG setting and those at the highest quality
setting. For example, if you take a picture with Canon EOS Rebel XS set to
its lowest resolution (1936 1288 pixels), the resulting JPEG image will be
around 700K. The same image taken with the camera set to mid-range
resolution (2816 1880 pixels) would yield a JPEG around 1.2MB. If you
were to use the highest JPEG setting (3888 2592 pixels), the image would
be around 3.8MB. Switching the high- resolution setting to RAW bumps the
file size up to almost 10MB. As you can see, it’s important to pay attention to
the settings you use. Memory cards are relatively inexpensive these days.
You need more than one. Keeping a camera set to the highest resolution and
best JPEG setting is normally a good choice, though using RAW offers
advanced users more flexibility. See Figure 8.14 in the “Graphics” chapter
for another example how changing the resolution and JPEG setting affects
image file size and the total number of images a memory card can store.
FLASHBACK
The Decisive Moment
For photographers who work on location, the two most important factors in image making are being there
Photographically, the decisive moment is that instant when an action is at its peak, when it’s stronger than
294 SECTION 3 Static Media
FIGURE 10.11
In this set of images, the photographer came upon a couple at a county fair who were
arguing. In the first of the 10 photographs (not shown) the young man is yelling at his
girlfriend, who’d disappeared for an hour. In the final frame, the couple kissed. The first
images held more aggression than desired, and the final seemed overtly sappy, so the 9th
image, which held the anticipation of making up, became the chosen photograph in the
sequence. Photographer: Chris Walker.
It took more than 100 years of photography for the term to be needed, but less
than half that for it to become obsolete. In the early years of photography, images
were exposed over a period of minutes—so timing was far less important than
the photographer’s location and composition were. Today we find that technology
—the force that, through faster shutter speeds, once created the need for
Bresson’s
statement—has brought us to a point in history where photographers no longer
need to exercise such decisiveness in their timing.
Through the ability to extract still images from video, we will soon be freed from the
necessity of precision timing, from the need to depress the shutter at the precise
moment an action is at its zenith. Almost as retribution, though, we must now further
challenge ourselves to be even more decisive in our selection of focal length, depth of
field, and composition.
by Chris Walker
Photography CHAPTER 10 295
EXPOSURE
A digital camera creates a picture by exposing the image sensor to light.
Underex- posed images have a dark appearance and lack detail while
overexposed images often contain bright spots and appear washed out. One of
the most important tasks of the photographer is managing the exposure
settings of the camera dur- ing image acquisition. Exposure is the product of
two variables: the intensity of light (as controlled by the iris), and the
duration of contact with the image sen- sor (as controlled by the speed of the
shutter). This relationship, the Reciprocity Law, states that there is an inverse
relationship between intensity and time and can be expressed mathematically
as:
Exposure Intensity Time
For example, setting the shutter speed to 1/250th and the aperture to f/8 results
in the same exposure level as a setting of 1/125th at f/11 or 1/500th at f/5.6. In
this example, doubling time while reducing the aperture by one full stop has
the same negligible effect on exposure as halving the time of exposure while
increasing the aperture by one full stop. As long as a change to one of the
two variables is matched by a reciprocal change in the other, exposure remains
constant.
FIGURE 10.12
Light bending as it
passes through a lens
or small opening is
nothing new—the
Chinese philosopher
Mo Ti first noted this
phenomenon 7,000
years ago when
viewing images of a
solar eclipse being
projected on the
ground through leaf
gaps in overhead
trees. So whether it’s
film, an image sensor,
your hand, or a white
card for viewing an
eclipse, the only magic
here lies in what we
do with this fantastic
medium. Photographer:
296 SECTION 3 Static Media
Chris Walker.
Photography CHAPTER 10 297
FIGURE 10.13
This image was shot
at 1/60th at f/22
(left), and at 1/1000th
at 5.6 (right). The
overall amount of
light striking the
camera’s sensor was
virtually the same in
each image, but the
slower exposure time
resulted in the image
with the blurred ball.
Photographer: Chris
Walker.
FIGURE 10.14
Correct exposure can
only be accomplished
in the camera, but
Photoshop and other
programs allow for
some adjustments
to be made in
postproduction. In
the center frame, the
Photoshop tool Curves
was accessed and
highlights and mid-
tones were elevated.
For the bottom image,
the History Brush was
used to return the
area surrounding the
lamp to its original
tone, then Curves was
again employed to
add some magenta to
the final color palate.
Photographer/Illustrator:
Chris Walker.
298 SECTION 3 Static Media
GREAT IDEAS
Flash Control
Using the flash allows you to provide additional fill light for scenes where existing lighting is less than ide
Fill flash (always on)—in this mode, the camera fires the flash every time, even when the exposure mete
Photography CHAPTER 10 299
FIGURE 10.15
Images made in direct sunlight can often benefit from a little judicious fill flash.
The objective of using fill flash is to “fill” in the shadowed areas that may be too
dark without it. Using too much fill, though, can unnaturally set your subject apart
from the background. A good starting point is to set your flash to –1. If you’re
using a
nondedicated unit, try setting the ISO to a higher setting—either way, the flash will
get the message that you desire less light on your scene than what it might want to
give you. Photographer: Chris Walker.
Auto Flash—in this mode, the camera meters the available light and only
fires the flash when needed. Limiting the use of the flash helps conserve battery
power.
Red-eye reduction—in this mode, the red eyes caused by the reflection of the
flash off of the retina of the eye is reduced. When the flash is set to red-eye
reduction mode, the camera fires a short burst of light followed by the actual
flash. The short burst is intended to shut down the pupil of the eye, thus
minimizing the kickback of reflected light. The closer the flash is to the lens,
the more prone a camera is to producing red-eye. On professional cameras,
the flash is farther away from the lens than on built-in systems, and can often
be detached and repositioned. In addition to controlling red-eye with the flash,
you may also be able to reduce it by using a wide-angle lens, increasing the
amount of existing lighting, or by moving the camera closer to the subject.
In a worst case scenario, you may be able to remove it using photo-editing
software.
WHITE BALANCE
Our perception of colored objects in the natural world is affected by the type
of light source used to illuminate them. Most light sources are not “true white”
but rather an uneven mix of the primary color frequencies (red, blue, and
green). For example, we know that under certain atmospheric conditions, the
physical color of reflected sunlight can appear red, orange, white, yellow, or
blue (see
300 SECTION 3 Static Media
Figure 10.16). In the same way, there are many different types of artificial light
sources (fluorescent, halogen, tungsten, etc.), each with its own unique color
properties. Light sources are rated according to their color temperature on the
Kelvin scale, so named for the British physicist William Thomson, 1st Baron
Kelvin of Largs (1824–1907) (see Table 10.4).
While the human eye adjusts rather easily to variations in light, electronic
image sensors must be calibrated to the color temperature of the light source
they are exposed to in order to accurately capture the color of objects within
the field of view. This process of calibration is called white balance. The
white balance of a camera can be set manually by shooting a white object
such as a blank sheet of paper while depressing the manual white balance
button. Once a camera “sees” what white looks like under existing light, it
can extrapolate the values of all of the other colors in the spectrum. Manual
white balance is the most accurate way of calibrating the color space of a
digital camera, especially when the source of light is mixed, such as an
indoor scene with natural light pouring in from an open window.
Digital cameras often provide white balance presets for several of the most
common lighting conditions, including daylight and cloudy for outdoor situa-
tions, and tungsten and fluorescent for indoors. However, if you forget to
change a white balance preset when moving from one light source to another,
the color
Table 10.4 The Color Temperature of Common Indoor and Outdoor Lighting Sources (Kelvin Un
FIGURE 10.16
This scene was
shot late in the day
under clear skies.
The blue could
have been natural
had the exposure
been
longer, but the longer
exposure would have
caused the drifting
fog to blur, so a bluish
filter was used to
emphasize the blue
light while allowing
an exposure time
that still showed
individual wisps of
fog. Photographer: Chris
Walker.
of your newly acquired images will not be true (see Figure 10.17). For this
rea- son, many people take the easy way out and just leave the camera set to auto
white balance (AWB).
By now, we’ve become used to the fact that digital cameras include
automatic settings for just about every control. While AWB works reasonably
well in most cases, you can usually get better results by using a white balance
preset based on what type of light you are working with or by manually white
balancing your camera.
FOCUS
The focus control is used to define the sharpness of an object within the frame by
changing the distance between the optical elements of a lens. A camera’s focus
must be reset each time the distance between the camera and the subject physi-
cally changes. Holding the camera steady while depressing the shutter button
will prevent unintentional blurring of the image upon capture.
Autofocus (AF)
When a camera is set to autofocus mode, it analyzes the area of a scene and
then calculates the distance between the camera and the main subject. Point-
and-shoot cameras normally use an active autofocus method in which an
invis- ible infrared beam scans objects located within 25 feet of the lens. The
camera
302 SECTION 3 Static Media
FIGURE 10.17
Quality lighting is vital to good photography. In this comparison, the sequence of three images on the left were made
at different times of the day; note that the brickwork on the building is reddest during the noon exposure, and shifts
as the day goes on due to the amount of blue, ambient light coming from the sky. In the images on the right (top to
bottom), the windsurfer was shot moments after dawn, during “the golden hour”; the woman at the wood rack was
shot at dusk, where you can see the sky light (blue) mixing with a mercury vapor light coming from the left (green)
and tungsten light from her patio and home (orange); the street festival image was balanced for the fluorescent
lights in the background, but it also contains “bug lights” within the corn booth (yellow), and high pressure sodium
lights from overhead (amber). Photographer: Chris Walker.
Photography CHAPTER 10 303
computes subject distance by analyzing the time it takes for the light to be
reflected back to the camera. DSLRs typically use a passive autofocus technique
that’s much more sophisticated and usually more accurate. With this
approach, the camera uses a small CCD sensor to detect and analyze changes
in the con- trast of image pixels. Active autofocusing systems can be used in
virtual darkness as they provide their own focusing light. Passive systems
must have a sufficient amount of light and image contrast in order to work
effectively.
AF TARGET POINT
By default, autofocus is usually set to a single target point located in the center
of the image. When the main subject is positioned in the center of the frame,
this works well. Simply set the focus and snap the image. However, a different
focusing strategy must be employed when the subject is placed in an off-
center position. With most cameras, this involves the following steps: (1)
place the main subject in the center of the screen; (2) acquire and lock the
focus by depressing the shut- ter button halfway; (3) reframe the main subject
in an off-center position; and
(4) fully depress the shutter button to capture the image. Many cameras provide
options for acquiring focus with user-select, single-spot, or multi-spot target
points.
While active and passive autofocusing systems have improved dramatically over
the years, neither method is foolproof. At some point, autofocus will let you
down as the camera miscalculates the proper focus setting. Problems are
likely to occur in the following situations:
■ When near and distant objects are mixed close together within the frame,
the camera may be unable to determine which object in the composition to
focus on.
■ Moving objects like cars and bikes can confuse the autofocus sensor, causing
a focus shift at the point of acquisition.
■ Extremely bright lights or subject areas can make it difficult for the camera to
lock onto a subject and acquire focus.
■ Monochromatic scenes can lack the contrast necessary for acquiring an accu-
rate focus.
GREAT IDEAS
Selective Focus
Photographs are two-dimensional. They only have width and height. We infer depth based on visual clu
background and/or foreground elements of the scene are blurred creating a sense of compositional de
DEPTH OF FIELD
Depth of field (DOF) refers to the area of a scene in front of and behind the
main subject that is in focus. The term great depth of field is used to describe
a photograph where the majority of the scene is sharply defined. Shallow
depth of field describes an image where noticeable portions of the foreground
and background areas of a scene are out of focus (see Figure 10.18).
Photographers will sometimes manipulate DOF to emphasize the focal point
of a composi- tion through visual contrast. Background elements can often
steal attention away from the main subject in a photograph. For example,
when an image is taken of a person positioned directly in front of a tree or
wall, the subject and the background appear to merge. The lack of three-
dimensional depth creates a flat image lacking in visual contrast. Decreasing
DOF in a scene such as this deemphasizes the visual importance of the
background by drawing the atten- tion of the viewer to the portion of the
scene that is in focus. In some cases, the photographer will use an
exceptional shallow depth of field, blurring the background to the point that
it may be all but unrecognizable. In photography, this effect is often referred to
as bokeh (BOH-ka), from a Japanese term meaning blur or haze.
The factors affecting DOF in an image are rather complex and would require
more time than we have to go into here. In the simplest terms, it can be said
that DOF is largely affected by three main variables: (1) the size of the lens
aperture or f-stop setting; (2) the focal length of the lens; and (3) the distance
from the camera to the subject.
Aperture Size
The size of the aperture is inversely related to the DOF of an image. As the size
of the aperture decreases, DOF increases, causing more of the scene to appear in
focus. As the aperture is enlarged, DOF decreases, creating greater contrast
between fore- ground and background objects. In a brightly lit scene, the
aperture is normally small, making it more difficult to acquire shallow depth of
field. In situations like this, a larger aperture setting can be obtained by
zooming in, or by increasing the camera’s shutter speed or decreasing the film
speed (see Figure 10.19).
Photography CHAPTER 10 305
FIGURE 10.18
These images were made almost identically, with the only changes being the shutter speed and aperture. The aperture
changes each in a way that alters its success through directing the viewer’s gaze. While neither is necessarily “right,” the
image with less depth of field seems more successful, since in the one with greater depth the decoy’s bill seems to be
competing for attention with the tools in the background. Photographer: Chris Walker.
FIGURE 10.19
Depth of field is the front-to-back distance within a scene that appears to be in focus because of the aperture at which
the image was made. In this comparison, the top image, shot at f/2.8, exhibits very shallow depth of field, while the
bottom image, shot at f/22, exhibits maximum depth of field. Illustrator: Adam Bulgatz.
306 SECTION 3 Static Media
Focal Length
The focal length of the lens is also inversely related to DOF. As you zoom in on
an image, the focal length of the lens increases while DOF decreases. For this
reason, wide-angle shots such as landscapes and vistas often have a great depth
of field, while narrow-angle shots acquired with a telephoto lens often have a
shallow depth of field.
Subject/Camera Distance
The physical distance between the camera and the subject is the third
variable affecting DOF in a composition. DOF increases with distance, and
decreases as you move the camera physically closer to the subject.
EXPOSURE MODES
Digital cameras feature a variety of exposure modes designed for many of
the most common shooting conditions. The exposure mode options vary by
cam- era, but they normally span a range from fully automatic, where control
of image exposure is retained entirely by the camera, to fully manual, where
the photographer assumes command of all exposure settings. A number of
semi- automatic modes exist in between the two extremes, allowing the
photographer and camera to share control of exposure settings.
Automatic Modes
FULLY AUTOMATIC MODE
When set to fully automatic mode, a camera will analyze the area of a scene
and calculate the settings for shutter speed, aperture, ISO, white balance,
focus, and flash. Depending on the quality of the camera, shooting in auto
mode can pro- duce good results under normal lighting conditions most of
the time. While shooting in auto mode makes the camera easier to operate, it
significantly reduces the photographer’s role in the creative process.
PORTRAIT MODE
In portrait mode, the exposure controls of the camera are optimized for shooting
close-ups with a shallow depth of field. The objective in this mode is to keep
the subject in focus while blurring the background. The camera accomplishes
this by raising the shutter speed in conjunction with widening the aperture.
Por- trait mode works best when the subject is tightly framed and there is
considerable distance between the subject and the background. Shallow DOF
is always more difficult to achieve when the subject is placed directly against
the background.
LANDSCAPE MODE
Landscape mode is the opposite of portrait mode. In this mode, the exposure set-
tings of the camera are optimized for shooting wide shots with a great depth
of field. The objective in landscape mode is to keep as much of the scene as
pos- sible in focus. To accomplish this, the camera combines a small aperture
Photography CHAPTER 10 307
with a
308 SECTION 3 Static Media
slower shutter speed. As a result, in some cases, you may need to use a tripod
to maintain a steady shot.
SPORTS MODE
Sports mode, or action mode as it’s sometimes called, favors a fast shutter speed
and is recommended for shooting a moving subject within a scene. When
cap- turing moving subjects during a bike race or soccer game for example, a
fast shutter speed is required in order to freeze the image without blurring. If
the shutter speed is too slow, a motion blur will occur. When composing an
action scene, a dramatic effect can be achieved by panning the camera along
with the moving subject while depressing the shutter button. This technique
produces a sharply defined subject while creating an intentional motion blur
in the back- ground (see Figure 10.20).
NIGHT MODE
Night mode uses a slow shutter speed combined with the flash when shooting
a subject set against a dimly lit backdrop like a sunset or evening sky. The
long exposure time allows the details of the background to remain correctly
exposed during the firing of the flash to illuminate the subject. A tripod
should be used in order to prevent unintentional blurring.
FIGURE 10.20
Sports mode is useful
for stopping action and
having enough depth
of field left to make
your images sharp. But
sometimes it’s good
to experiment and
“pan” a few frames.
Panning is the process
of selecting a slow
shutter speed, such as
1/8th or 1/15th, and
triggering the shutter
while following the
action. This method is
not for the impatient;
photographers who
succeed at panning
learn early on to shoot
hard and edit harder.
Photographer: Chris
Walker.
Photography CHAPTER 10 309
Tech Talk
Stitching Mode Stitching mode is used for acquir- mode, the camera provides a split screen view of the
ing a sequence of shots that can be joined together previous shot and the currently framed scene to help
in editing to create a large panoramic view of a scene you align the end of one frame with the beginning of
(see Figure 10.21). After taking the first picture in stitch the next.
FIGURE 10.21
Stitching mode can be used to take a series of photos that will be loaded into Photoshop, or an internal camera program,
to “stitch” them together. The key to success lies in overlapping your images—between a third and a half of the area
of the previous—so the computer can read the design of one image as it prepares to lace it together with the next
one in the series. This stitched image, a tiny courtroom where Lincoln practiced law, is a composite of six vertical
photos. Photographer: Chris Walker.
Semi-Automatic Modes
APERTURE PRIORITY MODE
In aperture priority mode, the photographer determines the f-stop setting manu-
ally, while the camera sets the remaining exposure variables. The main purpose
of using this mode is to control for depth of field in a scene. However, unlike
por- trait or landscape modes, the photographer retains total control of the
aperture.
310 SECTION 3 Static Media
FIGURE 10.23
The length of exposure
needed for this image
ultimately gives the
illusion of clouds or
fog mingling among
the rocks below. In
reality, though, it’s
the frothy heads of
crashing waves that
have been rendered
softly over the course
of the 20-minute
exposure, cumulatively
being gathered as
they blur and overlap
on a single piece of
film.
Photographer: Chris
Walker.
Photography CHAPTER 10 311
FIGURE 10.24
By opting for a lengthy
exposure, the water
was allowed plenty of
time to blur—about
a minute—while the
photographer’s luck
held out and very little
wind stirred the leaves
in the foreground.
Photographer: Chris
Walker.
PROGRAM MODE
Some cameras offer a program mode setting in which the camera determines the
correct settings for aperture and shutter speed, while allowing the photographer
access to other controls like ISO, white balance, flash, and so on.
Manual Mode
In full manual mode, the photographer has control of all of the camera’s
expo- sure settings and related controls. While this mode requires the most
skill and knowledge, it also allows for the greatest degree of creative control.
IMAGE STABILIZATION
Tripods, monopods, and other camera mounting devices are used in still pho-
tography and video production to keep the camera stable when shooting.
Even the slightest movement of the camera during acquisition can result in a
fuzzy image, or a distracting visual shift when using a video camera. The
potential for shakiness increases when the camera is set to a slow shutter
speed and/or to a long focal length. Therefore, it’s critical to stabilize the
camera in low-light conditions or in situations where you are not close to the
subject. In video pro- duction, shooting from a tripod should be considered
the norm rather than the exception. Random or unmotivated camera
movements are one of the telltale signs of an amateur production. While
using a tripod adds another degree of
312 SECTION 3 Static Media
difficulty to a production setup, having stable shots makes it worth the effort.
In situations where using a tripod is not possible, you can help the situation
by moving closer to the subject, adjusting the lens to a wide-angle setting,
and supplementing existing light, particularly if it is a bit dark.
Video cameras sometimes offer a digital image stabilization mode that works
by shifting the pixel matrix in the direction of camera movement in order to
counteract the effects of camera motion. While it may work well under
certain conditions, it is a digital effect and its use can lead to the creation of
unintended visual artifacts, making it a poor substitute for using a tripod.
GREAT IDEAS
Memory Cards
The images captured by a digital camera are usually stored on a removable flash memory card. These cards can best be thou
The CompactFlash (CF) card format was introduced by SanDisk Corporation in 1994 and features a footprint that’s a bit sm
FIGURE 10.25
For today’s young generation of digital photographers who’ve never shot with film, the idea of manually loading a spool of
(continued)
Photography CHAPTER 10 313
The Secure Digital (SD) card format was developed jointly by Matsushita, SanDisk, and Toshiba, and w
The MultiMediaCard (MMC) format was launched by Siemens AG and SanDisk in 1997. MMCs come in th
Sony developed the Memory Stick (MS) in 1998, as a proprietary format for their line of multimedia pr
DOWNLOADING IMAGES
Most cameras made today have either a USB or Firewire port built into the
cam- era for connecting directly to a computer to download images. Assuming
that all of the necessary hardware drivers are installed, your computer should
automati- cally recognize a camera connection as soon as the cable is attached,
the power is turned on, and the camera is placed in playback mode. In Windows
7, plugging in the camera should trigger the AutoPlay box, which will present
you with options for viewing, selecting, and downloading images from the
camera’s memory card to the computer hard drive. Unless you specify a
different location, Windows 7 will store downloaded images in the My Pictures
folder of the currently logged-in user. In Mac OS X, the default imaging
software is iPhoto. On a Mac, images are stored by default in the iPhoto
Library subfolder located in the Pictures folder of the current user. Both
operating systems also give you the option of manually moving your images
from the camera (or memory card) to the computer.
The AutoPlay wizard in Windows 7 is primarily a download interface and
contains very few controls for manipulating image content or file details. But
iPhoto is a bit more robust in that it can be used to perform a few basic
manip- ulations such as image rotation, red-eye correction, and simple color
correc- tion. It also provides some simple cataloging options for managing
the image library, resizing, and exporting images. Windows 7 provides some
of these same functions through the File and Folder Task Pane of Windows
Explorer.
photographer should take the time to: (1) properly assign a descriptive name
to each image file; (2) store related sets of images in topical subfolders; and
(3) modify the metadata fields of an image file to include details about the
picture’s origins and subject matter. While this may seem like busywork it is
not. Asset management is an incredibly important part of any multimedia
endeavor, and invariably leads to a better workflow.
RENAMING IMAGES
Digital cameras usually assign rather meaningless names to image files.
Renaming your image files is one of the first and most important things you
can do to bring a sense of order and uniformity to a collection of related
images. Programs such as Adobe Bridge and Apple’s Aperture can help
automate this process. When you rename the files, consider using a standard
prefix for related images. For example, for a group of photos taken of Beth on
her birthday in 2006, you could use beth_ bd06_ as the prefix. As we talk about
in Chapter 7, “Web Design,” use lowercase names if you plan to use the images
online. Similarly, use an underscore instead of a space when you want to
separate the character elements in a filename. After establishing the prefix,
you need to make the remaining part of the file names sequential. This is
where using a program to automate the process really helps. In some cases, you
may be able to use part of the file name generated by the camera. This is very
helpful for preserving the chronological order of the collection.
IMAGE FOLDERS
Try to avoid the habit of storing all of your digital images in a single master
folder. As your image library grows, the folder will swell, and it will become
increasingly difficult for you to navigate such a large collection of files in one
location. A better method is to use subfolders to store related sets of images.
Some of the same naming strategies we’ve discussed so far can be applied to the
naming of image subfolders. Sorting images into subfolders according to proj-
ect or client name, subject name, or event title, will help alleviate a great deal of
confusion and expedite locating image assets.
Tech Talk
Metadata Every time an image is captured and saved by
metadata using an international standard called the
a digital camera, nonpicture information about the image
Exchangeable Image File Format (EXIF). Nearly all image-
is recorded and saved to the file header. This informa-
editing programs are capable of displaying EXIF meta-
tion is called metadata and includes things like the
data. In Photoshop, selecting File>File Info opens up a
date and time of image acquisition, the camera make
dialog window containing access to EXIF metadata along
and model number, image size, and values for exposure
with a number of blank entry fields that the
set- tings like aperture, shutter speed, focal length, ISO,
photographer can use to attach more descriptive
and in some cases GPS data. Most cameras today
details to an image file (see Figure 10.26). Be careful
encode
about metadata. It
316 SECTION 3 Static Media
FIGURE 10.26
With each image, the camera gathers a significant amount of data that is available for retrieval long
after the photographs from a shoot have been made, edited, retouched, and cataloged. In addition to a
complete list of camera settings, some models even gather GPS data, allowing photographers to later
identify locations or revisit them for subsequent shoots. Photographer: Chris Walker.
Photography CHAPTER 10 317
CHAPTER SUMMARY
A digital camera is one of many tools a multimedia producer can use to create
visual assets for a project. The more knowledgeable you are about the imaging
process, the more creative freedom you will be able to exercise during the
acquisition of digital pic- tures. This chapter has focused primarily on the mechanical
aspects of digital photogra- phy by discussing the operational function of imaging
components and the management of exposure controls. As the photographer, the degree
to which you retain control during the exposure process is determined by which shooting
method or mode you adopt. In fully- automatic mode, the camera retains total control
of all the exposure variables. In manual mode, the photographer has more latitude and
creative freedom but must also possess greater skill and knowledge about the camera,
its operation, and intentional outcomes.
This page intentionally left blank
SECTION 4
Time-Based Media
Yuliyan Velchev/Shutterstock.com
This page intentionally left blank
CHAPTER 11
Audio Production
315
Key Terms
3-Band Equalizer A/D Converter A
The notion of media will be completely blended in one digital stream. AGC AIFF
Moving pictures and static pictures and text and audio—the mix of Ambient Sound Amplitude Audio
Bass Roll-Off Bidirectional Bit Dep
any of those will be commonplace. Imagine turning the pages of a Microphone Container Format De
magazine: every one of the images is a sound clip, and when you put EQ
your cursor over one of them, the audio comes. That’s absolutely going Frequency Gain Handheld
Microphone Headphones
to be happening in 15 years. Full convergence of media and High-definition Audio Hypercardio
television and computers. Mini Plug Monitoring MP3
—Jim Clark, Computer Scientist and Founder of Silicon Graphics Inc. and
Netscape, (1994)
Chapter Highlights
This chapter examines:
The nature of sound and audio
The audio chain and signal flow
Microphone element designs, pickup patterns, and form factors
Microphone placement and recording techniques
Audio cables, connectors, and cable management
the images
Omnidirectional Over-Under and sound
Wrap Phantom Powerof a live-event are often acquired at the same moment,
Pitch
Plosives Polar Pattern Preamp
the workflows for recording sound for use in a preproduced video or film are
Proximity Effect Pulse-Code
not necessarily the same, and it is important to understand the differences.
Modulation (PCM) Quantization Rarefaction
This chapter
RCA Connector Ribbon Microphone focuses
Sample on the basic concepts, tools, and techniques you need to
Rate Sampling
Shotgun Microphone SOTbe aware of as you delve into sound acquisition, recording, and editing.
Sound Sound Check
Sound Pressure Level (SPL) Sound Pressure
WHAT IS SOUND?
Wave Supercardioid Transduction Treble
TS/TRS Connector Ultracardioid Unbalanced Audio XLR Connector
VU Meter WAV It’s important to distinguish between the physical characteristics of sound as
a phenomenon of nature, and the field of audio production—the capture and
reproduction of sound. Sound is a natural phenomenon that involves pressure
and vibration (see Figure 11.1). Understanding how sound and hearing work will
help you capture and produce better quality audio. What we perceive as
sound traveling across time and distance is actually the invisible moving
presence of a sound pressure wave. Sound waves are a special type of energy
and require a molecular medium for propagation. They can travel though solids,
gases, or liq- uids, but air molecules are the best transmission medium.
Despite Hollywood’s portrayal of loud space battles, sound cannot travel
within the vacuum of space, and in reality, a ship exploding in space would
not make a sound as there are no air molecules. We hear by discerning changes
in the pressure and movement of the air particles around us. When a tree falls
in the forest, the air molecules in its path are momentarily displaced. They
are violently pushed out of the way to make room for the falling tree. This sets
off a chain reaction as the energy of the
FIGURE 11.1
Striking a tuning fork
causes its two prongs
to vibrate, which
in turn, produces a
musical tone.
Audio Production CHAPTER 11 317
initial force is passed along to neighboring molecules in all directions. Back and
forth they go, oscillating until the energy that caused the disturbance
dissipates.
The best way to illustrate the movement of a sound pressure wave is to look
at something more visible in nature. When you drop a small rock into a still
pond, you see concentric ripples or waves traverse outward from the point of the
disrup- tion (see Figure 11.2). Here, water serves as the conduit for energy to
flow away from the source. However, the actual water molecules travel only a
tiny distance as they bounce back and forth transmitting the energy signature of
the wave. As the wave travels further away from the source, the oscillations
of the molecules begin to slow down until the pond is once again at rest.
FIGURE 11.2
Concentric ripples are
the visible evidence of
molecular vibration.
Uniform waves
progress outward in
every direction, as
energy is released
at the point of
disturbance.
Tech Talk
Amplitude
The first thing we tend to notice about sound is how loud it is. Loud sounds
capture our attention almost immediately, while soft sounds strain our senses
or elude us entirely. Because sound waves are invisible to the human eye, we
must use pictures to illustrate their physical qualities and characteristics (see
Figure 11.3). A sound wave’s height (amplitude) indicates the intensity or
magnitude of the pressure wave. Amplitude is defined as the distance from the
crest of the wave to the trough. The louder the sound, the greater the amplitude,
and the taller its waveform will be. The amplitude of a sound is greatest near the
source and diminishes over distance and time.
Amplitude is measured in decibel units. The decibel (dB) is a logarithmic
unit of measurement used to quantify the sound pressure level (SPL) or
magnitude of a sound wave. Humans are capable of hearing a wide range of
sounds, from 0 dB to 120 dB. A value of 0 dB represents the least audible
sounds that we can hear(just above silence). With each increase of 20 dB on
the decibel scale, the amplitude and perceived loudness of sound increases 10
times. Thus, a 20 dB sound pressure wave is 10 times louder than the faintest
sound. A 40 dB sound source is 100 times louder than the faintest sound and
10 times louder than sound at 20 dB. When you reach 120 dB, the SPL is
1,000,000 times greater than the level of sound at 0 dB. While adapting to a
logarithmic scale can be confusing at first, using a scale with a relatively
small number of decibel units is easier to deal with than one with a million or
more increments of variation.
The human threshold for pain begins around 140 dB, while permanent
hearing loss occurs at 150 dB. Hearing loss most commonly results from
repeated expo- sure to loud sounds over time. With the growing popularity of
digital music and MP3 players, concern has emerged over the potentially
damaging effects of listening repeatedly to loud music through headphones
or earbuds. The close proximity of earbuds to the sensitive organs of the ear
makes this an even greater concern and has prompted the makers of personal
listening devices to
B C
A
Time
1 second
FIGURE 11.3
Sine waves are often used to visualize the repetitive oscillations of sound vibrations.
(A)Amplitude is represented by the height of the wave. (B) Wavelength is the distance traveled
during one complete vibration cycle. (C) Frequency is the number of complete wave cycles
that occur over a set period of time (usually measured in 1-second intervals).
Audio Production CHAPTER 11 319
Tech Talk
offer volume-limit controls on their units. A volume limiter option allows users
to set a maximum listening level based on decibels or the relative volume
units of the device. While many factors can contribute to hearing loss or
damage, setting a volume limit in place is advisable. By the way, that ringing
you get in your ears after a concert—it’s called tinnitus and can become
permanent. Trust us, it’s not fun.
Frequency
As sound waves pass through matter, the vibrating molecules experience
three phases of movement (see Figure 11.4). As molecules move in an
inward direc- tion, they are pushed closer together, leading to an increase in
molecular den- sity and sound pressure. This is the compression phase and is
represented by the portion of the waveform above the horizontal axis (time).
The highest point of the waveform is called the crest and signifies the
moment of greatest sound pressure. Once maximum compression has been
reached, elasticity kicks in, causing the molecules to return to their original
position. For a fraction of a second, the molecules are at rest as they change
direction and begin moving outward. During the rarefaction phase,
molecules are pulled apart, resulting in a decrease in molecular density and
sound pressure. Rarefaction is denoted as the portion of the waveform below
the horizontal axis. The lowest point on the waveform is called the trough
and indicates the moment of lowest sound pressure.
The progression of a sound wave through one phase of rest, compression, and
rarefaction is called a cycle, and a sound’s frequency is based on its number
of cycles-per-second. Frequency describes a sound’s relative low or high pitch.
Frequency is measured in hertz (Hz), cycles per second. Every vibration has a
unique frequency signature. A common frequency used in audio production
for the purposes of calibration is the 1 kHz tone. By the way, 1 kHz is simply
1,000 Hz. Kilohertz (kHz) units can be used as an abbreviated way of referring
to particularly high frequencies.
When individually plucked, the strings of an acoustic guitar in standard
tuning create the frequencies, or pitch, shown in Table 11.1. Each string can
be played separately, producing a single note or pitch. However, most of the
time a musician will strike multiple strings at a time, thus creating diverse
sounds and harmon- ics with complex frequency signatures. Most sounds in
nature, including human
320 SECTION 4 Time-Based Media
Compression (maximum pressure)—the density of molecules increases as they are pushed together
FIGURE 11.4
When the air
Sound Pressure (Amplitude)
molecules around us
are energized by a
sound pressure wave,
they begin to oscillate,
Time
bouncing rapidly
back and forth in
unison. As molecules
travel inward they
are squeezed
tightly together
(compression); as
they spring back
in the opposite
direction, they pull
apart and spread out
(rarefaction). At one
brief moment in each Rarefaction (minimum pressure)—the density of molecules decreases as they are pulled apart
wave cycle, when
molecules change
direction, they are
briefly at rest.
frequencies include lower pitch sounds in the range of 20–320 Hz. Midrange
frequencies include medium pitch sounds falling between 320 and 5,120 Hz.
Treble frequencies, high pitch sounds from 5,120 to 20,000 Hz, represent the
largest segment of the frequency spectrum.
GREAT IDEAS
Equalization
Sound systems often allow you to adjust the bass, midrange, and treble output of a program source or channel. This featur
and provides the user with separate controls for raising or lowering the gain of each frequency region or band (see ). For exa
A 3-band equalizer is simple and inexpensive, but it provides only global controls for adjusting the pitch of the recording or
FIGURE 11.5
A 3-band equalizer, like the one pictured here (top), is a standard feature on most audio mixing consoles and on many electr
322 SECTION 4 Time-Based Media
Pulse-Code Modulation
In order for sound to be understood by a computer, it must be converted from
its native analog format to a digital format that can be represented with
discrete numerical values. Pulse-code modulation (PCM) is the most
common method used for representing audio as a digital stream of binary
data. British engineer Alec Reeves invented the technique in 1937 as a way
of reducing noise dur- ing the transmission of speech over long-distance
telephone lines. The PCM method relies on a technique called sampling to
reduce a continuous signal into a sequence of discrete samples that can be
defined mathematically. Each unit (or sample) represents a measurement of the
sound wave’s properties at a single moment in time (see Figure 11.7). Each
measurement is recorded numerically as a binary string of 0s and 1s. The fidelity
of a PCM audio stream is determined by two basic properties: the sampling
rate, which indicates the number of samples that are recorded every second;
and the bit depth, which refers to the number of bits used to encode the value
of a given sample.
FIGURE 11.6
Audio recording
technologies have
changed often over the
years. Physical media
like vinyl records
that you could hold
and touch have been
replaced with invisible
data files that can
be purchased online
and downloaded to a
portable MP3 player in
a matter of seconds.
Audio Production CHAPTER 11 323
Analog Signal
FIGURE 11.7
An analog audio signal
is continuous without
interruption (top).
Sampling the analog
signal at regular
intervals is the first
step in producing
a digital recording
Digital Sampling (bottom).
Rate)
Audio sampling rates are designated in kilohertz (kHz) units. For example, a 48
kHz sampling rate indicates that the recording was sampled 48,000 times per
second. While working for Bell Laboratories, engineer Henry Nyquist
discovered that the sampling rate must be at least twice as large as the bandwidth
(or frequency range) of the signal being recorded. Since human hearing has a
frequency range of 20 Hz to 20 kHz, a minimum sampling rate of 40 kHz is
required in order to effectively capture and encode all frequencies within this
range. If the sampling rate is any lower, then frequency data (within the human
hearing range) may be lost during the recording process.
Bit Depth
Whereas the sampling rate directly impacts the frequency response of a digital
recording, bit depth affects its dynamic range and the amount of noise and
distortion introduced during the recording process. Each audio sample is stored
as a binary sequence of 0s and 1s. The bit depth setting determines how many 0s
and 1s are used to represent the value of a waveform’s voltage signature (or
amplitude) at each sam- pling interval. The more bits you have to work with, the
more descriptive the encod- ing can be to record subtle fluctuations in the
waveform’s amplitude. Quantization is a term that’s often associated with bit
depth, and is used to describe the mathemati- cal process of assigning discrete
324 SECTION 4 Time-Based Media
numerical values to each voltage measurement.
Audio Production CHAPTER 11 325
An 8-bit binary sequence has 256 possibilities. A 16-bit sequence has 65,536
pos- sibilities. And a 24-bit sequence can be arranged 16,777,216 different ways.
The most common bit depth settings used in professional audio and video
recording are 16-bit and 24-bit. While 24-bit recording is technically superior in
terms of dynamic range, the higher bit depth also results in larger audio files.
The addi- tional data will demand higher performance from your computer
system and editing software, potentially causing it to slow down if your system
isn’t powerful enough to handle 24-bit audio processing. While the difference
between 24-bit
FLASHBACK
Industry Standards
Sony and Philips are familiar names in the consumer electronics industry. They joined
forces to create the audio compact disc format in 1980. In doing so, they developed
the infamous “red book” standard for the audio compact disc (CD), which specifies a
bit depth of 16-bits and a sampling rate of 44.1 kHz. To this day, this is the industry
standard used when distributing digitally recorded music to consumers. No matter what
the bit rate and sampling rate are during production and editing, the final distribution
file for your audio project should conform to 16-bit/44.1 kHz before being burned to a
CD. In digital film and video production, a slightly higher sampling rate of 48 kHz was
adopted as the industry standard. This rate conforms to the encoding standards set
by Philips, Sony, Toshiba, and Time-Warner for the digital versatile disc (or DVD)
recording platform in 1995. Because of this, camcorders are often programmed to
encode the sound portion of a video recording at a sample rate of 48 kHz.
Since the introduction of HD television and Blu-Ray DVDs the concept of high-definition
audio has grown more popular. The HD audio standard conforms to 24-bit/96 kHz
encoding of the audio stream. While this standard is not yet universally supported,
the added bandwidth is necessary for a number of advanced surround sound and
multichannel applications. If your recording hardware supports 24-bit/96kHz encoding,
then by all means choose this setting. Acquiring at the highest resolution possible is
always a good idea. After your program is edited, you will have the option of exporting
it as a standard definition product (16-bit/44.1 kHz).
FIGURE 11.8
The recording format, sample
rate, and bit depth settings
are often located in the menu
system of the recording
device. Before you start
recording, make sure that
you review these settings.
As a general rule, choose
the highest quality setting
that your device supports.
You can always downsample
to a lower resolution later in
postproduction.
326 SECTION 4 Time-Based Media
and 16-bit looks relatively small, at a sampling rate of 44,100 kHz, the amount of
digital data stored every second becomes quite substantial. In a 16-bit/44.1
kHz recording, 700,600 bits of data are produced every second. In a 24-bit/96
kHz recording, the audio stream will contain more than 2.3 million bits per
second. You may hear people refer to audio encoded at 24-bit/96 kHz or greater
as high- definition audio because of its superior resolution.
FIGURE 11.9
In Apple iTunes, users
can convert audio
sound files into any
one of five formats:
AAC, AIFF, Apple
Lossless, MP3, and
WAV.
largely interchangeable, and virtually all sound editing programs provide support
for both of these uncompressed file formats (see Figure 11.9).
SOUND RECORDING
There are three main components to a sound recording system—a source, some-
thing or someone creating a sound; a microphone to convert physical sound
waves into an analog signal transmitted via an electric current; and a
recording device to store the sound imprint.
Microphones
A microphone is a recording instrument used to convert sound waves into an
electrical equivalent that can be stored, transmitted, and played back through
an audio sound system. Some microphones are designed specifically for use
in a studio environment, while others are optimized for field use. Likewise,
some microphones are better for voice work while others are designed
primarily for instrument recording. While there’s often more than one clear-
cut choice, it’s important to understand the fundamental differences in
microphone design in order to choose the most appropriate tool for the job.
FIGURE 11.10
A boom arm and
shockmount are
used to support the
microphone during
this recording studio
session. A pop filter
is placed in front of
the microphone to
reduce plosives, vocal
artifacts like those
infamous “popping P
s.”
sound wave. And like the original sound wave, analog recordings provide a con-
tinuous and uninterrupted representation of the original sound.
Microphones are generally classified according to three main features or charac-
teristics: (1) Transduction Method, (2) Polar Pattern, and (3) Form Factor.
DYNAMIC MICROPHONES
Dynamic microphones use acoustical energy and mechanical vibration as the
means for producing the electromagnetic signal required for analog recording.
Dynamic microphones do not require a power source. They are durable, relatively
inexpensive, and moisture- and shock-resistant. Moving-coil and ribbon
microphones are two of the most common types of dynamic microphones. Both
rely on electromagnetic induction, which uses magnets to produce an electric
current (see Figure 11.11-A).
MOVING-COIL MICROPHONES
In a moving-coil microphone, a diaphragm is attached to a coil (a metal core
wrapped with copper wire) suspended in a magnetic field between the north and
south poles of a fixed magnet. The diaphragm is a thin, circular membrane,
typically made of paper, plastic, or metal. As the diaphragm vibrates, the coil
oscillates in the magnetic field, producing a tiny current that’s transmitted via
copper wire to the microphone cable. The electromagnetic signal modulates in
330 SECTION 4 Time-Based Media
unison with the amplitude and fre- quency of the sound pressure wave,
producing a copy of the original waveform.
Audio Production CHAPTER 11 331
FIGURE 11.11
A) A dynamic
microphone is
less sensitive to
sound because
the transducer is
self-powered by
the sound of the
subject’s voice. B) A
condenser microphone
is more sensitive
to sound because
the transducer is
powered by a battery
or phantom power
source.
RIBBON MICROPHONES
A ribbon microphone uses a thin metal ribbon of corrugated metal, usually
alumi- num, as the transduction element. The ribbon is suspended in a magnetic
field between the opposite poles of a fixed magnet and generates an
electromagnetic current when it pulsates in the magnetic field. Ribbon
microphones are techni- cally superior to moving-coil designs because they
respond to sound bidirection- ally, from both the front and the back of the
element. While ribbon microphones are relatively expensive, broadcasting and
recording professionals value them for their superior performance and natural
sound reproduction. The metal elements in early ribbon microphones were
quite delicate, and ribbon microphones had a reputation for being easy to
damage. Newer ribbon microphones are more robust, though as with their
predecessors, you need to be careful about picking up wind noise when using
them outdoors (see Figure 11.12).
CONDENSER MICROPHONES
Condenser microphones use a capacitor to record variations in amplitude and
fre- quency. The capacitor has two parts, the back plate (containing the electric
charge) and the diaphragm. As the diaphragm vibrates, the distance between
it and the
332 SECTION 4 Time-Based Media
FIGURE 11.12
Vintage radio
microphones like
this one often have a
ribbon transducer.
back plate changes, thus modulating the intensity of the voltage signal.
Condenser microphones are much more sensitive to sound than dynamic
microphones, and as a result can be positioned further from the source of the
sound. Condenser micro- phones are separated into two groups based on
diaphragm size. Large diaphragm condensers have a bigger form factor and are
more often used in a studio recording environment, while small diaphragm
condensers have a slender body profile and may be found in both field and
studio environments (see Figure 11.11-B).
Condenser elements require an external power source to supply the electric
charge to the back plate. For this reason, condensers are often equipped with an
attached bat- tery pack or built-in power module. A single AA battery is usually
all that’s required. As an alternative, condensers can receive phantom power
directly from a connected mixer or recording device. Phantom power supplies a
48-volt (+48V) electric cur- rent to the capacitor through the attached
microphone cable. Professional audio mixers and video cameras usually include
this feature (see Figure 11.13).
FIGURE 11.13
The two microphones on the left are condensers. However, the one pictured at the top can be powered with either a
battery or phantom power. The microphone on the bottom does not have a battery compartment and must be powered by
the camera or recorder it is connected to. Most professional recording equipment can provide phantom power, but it must
be turned on to work. The phantom power switch may be located on the outside of the unit or, as shown on the right,
within the menu system of the device.
short, the polar pattern of a microphone impacts how you use it and under
which circumstances the microphone will function at its best (see Figure 11.14).
OMNIDIRECTIONAL
The pickup pattern of an omnidirectional microphone is a sphere around the
micro- phone, though not an entirely perfect one. In theory, these
microphones respond equally to sound in all directions. In practice, however,
the microphone body, par- ticularly on handheld microphones, can block or
obscure the path of a sound wave. This can shield the microphone a bit from
some frequencies. The smaller the micro- phone’s body, the less of a problem
this is. Because they pick up sound from all around, omnidirectional
microphones are best used in situations where there is lit- tle to no ambient
sound. You may also hear these microphones called nondirectional.
Audio Production CHAPTER 11 335
FIGURE 11.14
Six of the most
common polar patters.
(A)Omnidirectional.
(B) Bidirectional.
(C)Cardioid.
(D) Supercardioid.
(E)Hypercardioid.
(F)Ultracardioid
(or shotgun).
BIDIRECTIONAL
Bidirectional microphones pick up sound equally from the front and rear of the
ele- ment. Most ribbon microphones are bidirectional. As a broadcast
performance microphone, these are ideal for interviews where the host and
guest are seated on opposite sides of a table or in situations where two people
are required to share a single microphone.
336 SECTION 4 Time-Based Media
CARDIOID (UNIDIRECTIONAL)
As the name implies, a unidirectional microphone picks up sound from only one
direction. This makes it well suited for working in situations with lots of ambient
(background) noise. There are a number of variants of this type of
microphone. Cardioid microphones have a unidirectional polar pattern with a
heart-like shape (hence their name). This pickup pattern favors sounds coming
from the front and sides up to 130°. Cardioid microphones boast a relatively
narrow pickup field and do a good job of rejecting ambient sound from the
rear of the microphone. Cardioid microphones are ideal for recording single
subjects and vocalists. Other members of the unidirectional family include
supercardioid, hypercardioid, and ultracardioid (or shotgun) microphones. Each
progression comes with a narrower pickup field and an expanded area of sound
rejection from the rear of the micro- phone. The narrower the pickup pattern,
the more proactive the operator needs to be in aiming the microphone
directly at the sound source during recording.
HANDHELD MICROPHONES
Handheld microphones are designed for the talent or performer to hold during a
recording session. Dynamic handheld microphones are ideal for rugged use and
heavy handling, but they need to be held close to the mouth (3–6 inches) in
order to generate enough sound pressure for a good recording (see Figure 11.15).
The rugged construction of dynamic handheld microphones minimizes noise
caused by sudden movement, rough handling, or when passing the
microphone along from person to person. Reporters rely on this type of
microphone most often when recording a standup interview or conducting field
interviews. If you are interviewing someone using a directional handheld
microphone, remember that you need to angle it toward the other person
when he or she is talking.
FIGURE 11.15
Reporters often use
handheld microphones
for conducting field
interviews. Because
handhelds are usually
dynamic, they should
be positioned a few
inches in front of the
subject’s mouth.
Audio Production CHAPTER 11 337
Think about it as sharing an ice cream cone. If the other person is going to
eat, you’ve got to put the cone in front of him or her.
Some handhelds are equipped with an electret condenser element. While this
increases the microphone’s sensitivity, it raises the risk of unwanted noise
from outside forces. To reduce handling noise, condenser handheld
microphones usually come with an internal shock mount. The shock mount
suspends the element in mid-air with elastic bands, insulating it against sudden
jarring move- ments (refer back to Figure 11.10).
Most handheld microphones are unidirectional. Handhelds work best when
the microphone is positioned no further than 6 inches away from the mouth
and slightly off-axis. Getting too close to the microphone grill (the shield)
can pro- duce annoying artifacts such as the infamous “popping Ps” and
other unpleas- ant plosives—the sound caused by releasing blocked air in
speech, particularly from the pronunciation of the hard consonants b, d, g, k,
p, and t. Remember, the microphone element is highly sensitive to sound
pressure vibrations, and when a plosive ignites less than an inch away, vocal
distortion and clipping is likely to occur. Using a windscreen or pop filter can
help reduce or eliminate these types of negative consequences, but you still
need to pay attention to your distance from the microphone (see Figure
11.16).
FIGURE 11.16
When using any
microphone outdoors,
it’s a good idea to
attach a windscreen.
338 SECTION 4 Time-Based Media
Tech Talk
LAVALIERE MICROPHONES
Lavaliere microphones are designed specifically for speech and vocal
acquisition. Also known as a lapel or lav microphone, this low-profile workhorse
is designed with an electret condenser transducer element. Although they
come in a variety of polar patterns, the most common ones have an
omnidirectional, cardioid, or supercardioid element. Like all condensers, lavs
require a battery pack or phan- tom power signal from the mixer or recording
device.
Lavalier microphones have an extremely small form factor. Unlike handheld
microphones, lavs are designed to be worn and are attached to the talent’s
clothing—often a tie, collar, or lapel—with the aid of a tiny metal clip. These
microphones are too small to be handheld, as handling noise will produce
unwanted artifacts during recording. Since they are physically attached to the tal-
ent, the distance from the microphone to the source is constant, removing an
important variable from the recording process. Whether the subject is
running, walking, sitting, or standing, the position of a lavalier microphone,
relative to the source, will change very little. However, even when a lavaliere is
properly attached, you have to be careful. The talent’s physical actions (moving
hands, feet, clothing, etc.) can cause unwanted noise if the microphone is
suddenly bumped or jostled. Lavs are particularly popular with platform
speakers who want the freedom to walk and talk at the same time without
having to hold a microphone or stand behind a podium. They are also a good
choice for recording interviews, especially when used indoors in a quiet setting
(like a TV studio or office) and where the talent’s physical movements won’t
interfere with a good recording.
Proper microphone placement is essential for maximizing performance and
opti- mizing the rejection of unwanted ambient sound. Lavalier microphones
should be attached 3–6 inches below the neckline. The microphone element
should face upward towards the chin and be free of any obstructions from
clothing and jew- elry. Attaching a windscreen to the microphone capsule can
help to alleviate wind or breathing noise. Lavs should be positioned so that the
subject’s voice projects directly over the top of the microphone. Most of the
time, this means affixing it directly in the center of the upper body or chest.
However, if the subject is turned off-axis to the camera, then the microphone
should be positioned slightly to the left or right of center, so that it remains
directly under the mouth or chin of the subject.
To maintain a professional appearance, lavalier microphones should be properly
dressed. Dressing a microphone involves making it as attractive and obscure
as possible (see Figure 11.18). At all costs, be sure to avoid the rookie
mistake of allowing the microphone cable to dangle down the front of the
subject’s shirt or blouse. This is a telltale sign of an amateur production.
With just a little bit of effort, and discretion, the microphone cable can be
rerouted out of sight beneath clothing, or hidden behind a tie, lapel, jacket, or
collar. Discretion is critical, because working with people in a professional
production setting requires sensi- tivity to cultural norms and rules of etiquette.
While you need to keep these issues in mind whenever you are working, you
need to be particularly cognizant of these issues when working with people from
340 SECTION 4 Time-Based Media
other cultures or with different genders.
Audio Production CHAPTER 11 341
FIGURE 11.18
The lavalier is
improperly attached
and dressed in the
top two photos. This is
the mark of an
amateur production.
Taking the time to
properly attach and
dress a lavalier
microphone, as in
the bottom photos,
improves the quality
of the recording and
the appearance of the
subject during on-
camera interviews.
SHOTGUN MICROPHONES
Shotgun microphones are among the most directional microphones. They feature
a condenser element with an extremely narrow pickup pattern—in supercar-
dioid, hypercardioid, and ultracardioid varieties. They are so named because
of their long and slender form factor, which resembles the general shape of
342 SECTION 4 Time-Based Media
FIGURE 11.19
A boom pole is used
to position a shotgun
microphone within
a few feet of the
subjects in this
scene. The
videographer works
with the
boom pole operator
to ensure that the
microphone does not
dip down into the
visible portion of the
frame.
FIGURE 11.20
A pistol grip is used in
lieu of a boom pole for
acquiring the sound of
footsteps on leaves.
When using a shotgun
microphone to acquire
sound effects, the
audio engineer usually
wants to get the
microphone as close
to the source of the
sound as possible.
FIGURE 11.21
Different techniques
are used for
positioning a
microphone with
a boom pole. Most
often, the pole is held
above the subjects or
below them. Placing
the microphone too
close to the actors
(bottom photo)
may disrupt their
concentration and
potentially ruin a take.
344 SECTION 4 Time-Based Media
BOUNDARY MICROPHONES
Boundary microphones, also known as pressure zone microphones (PZM), are
condenser microphones intended to be placed on a flat surface, usually a
table, ceiling, or wall. As you can see in the illustration, the microphone is
mounted to a metal plate. The microphone points down at a metal base-
plate and detects the sound pressure waves in between the microphone and
the plate—the pressure zone (see Figure 11.22). This minimizes the effects
of sound reflecting off other sources such as a wall by limiting the sound
capture area. PZMs are usually used to record meetings and conferences and
are good for recording multiple people at the same time. As you can see,
they don’t really look like what we typically think a microphone looks like.
You can use this to your advantage—PZMs are less likely to make people
nervous.
FIGURE 11.22
The Audio-Technica
ES961 is a cardioid
condenser boundary
microphone designed
for use on a flat
surface such as a
conference table or
floor.
Source: audio-technica
.com
Audio Production CHAPTER 11 345
FIGURE 11.23
In the photo on the GREAT IDEAS
left, the producer uses Why Should You Use an External Microphone?
the Here are four reasons for using a professional-grade external microphone instead of one that is built int
built-in It gives you the ability to select the right type of microphone for the intended sub- ject and recording app
on his digital audio It gives you maximum control over microphone placement and distance to the subject, independent of w
recorder. In the photo All things being equal, a professional external microphone will have better sound fidelity than the built-i
on the right, he uses It can help you avoid tape noise on tape-based equipment.
the same recorder, but
attaches an external
handheld microphone.
All things being equal,
you will achieve
better results using a
professional external
microphone to conduct
interviews.
346 SECTION 4 Time-Based Media
Wireless Microphones
Although most of the time you are working with audio you’ll be running a wire
directly from the microphone to your recorder or mixer, you’ll occasionally
need to use a wireless microphone, particularly if you have talent who likes to
walk around the room a lot. In some cases, you can plug your regular micro-
phone into a wireless transmitter—essentially a tiny radio station that will send
the audio to a radio receiver—which you’ll plug into your mixer or recorder (see
Figure 11.24). You may lose some of your frequency range when you switch to
wireless. Like all battery-operated equipment, make sure you keep fresh batteries
on hand for the transmitter and the receiver. If you use more than one wireless
microphone at a time, you’ll have to make sure they aren’t using the same radio
frequency. Just as two radio stations can’t broadcast on 88.1 FM at the same
time, you can’t have two wireless microphones operating on the same radio
frequency. The better the system, the more frequency options you’ll have.
Remember, you may not be the only one using a wireless microphone at some
locations.
FIGURE 11.24
A wireless microphone
like this one is
designed for remote
field production.
The microphone
and transmitter
are attached to
the
subject. The receiver is
attached to the camera.
Source: audio-technica
Audio Production CHAPTER 11 347
.com
348 SECTION 4 Time-Based Media
There are two basic transmitter types for professional equipment, UHV and
VHF. Of the two options, UHF offers the most available frequencies. This may
be help- ful if you are using more than five wireless microphones at a time or
are going to be in locations where others are using wireless microphones. On
the other hand, UHF microphones are typically more expensive than VHF
microphones. Be aware that some of the radio frequencies overlap with
television station fre- quencies. Always test your equipment before you begin
production. A third option for wireless is an infrared transmitter, which uses a
beam of infrared light to send the signal. It’s usually best to avoid infrared
systems, as they require line- of-site transmission. If something gets between
the transmitter and the receiver, it won’t work, much like the remote control
for your television. Infrared does have the advantage of being more secure—
you can’t pick it up from outside the room—and infrared transmitters don’t
typically interfere with each other.
Digital Microphones
One of the challenges with the move towards a digital workflow involves record-
ing audio directly into the computer. While you can plug some microphones
directly into a computer, a better solution is to use an A/D, an analog-to-
digital converter, which translates the analog input into a digital format.
These often rely on Firewire or USB connections. Using an A/D isolates the
conversion pro- cess from the rest of the computer and lessons the chance of
computer compo- nents such as the power supply introducing unwanted
noise. Another option is to use a digital microphone—a microphone with the
A/D built into it. These microphones are often called USB microphones
because they typically connect to the computer via a USB connector (see
Figure 11.25).
FIGURE 11.25
A student uses a USB
microphone directly
connected to the
computer to record a
voiceover.
Audio Production CHAPTER 11 349
AUDIO CONNECTORS
Microphones may be connected to audio recording devices by a variety of cables
and connectors. Professional microphones use balanced connectors, while
con- sumer devices use unbalanced connectors.
FIGURE 11.26
Professional
microphones and
recording systems
have balanced XLR
connectors.
350 SECTION 4 Time-Based Media
XLR CONNECTORS
Balanced microphone cables usually have XLR connectors on each end of
the wire. The female XLR connector attaches directly to the end of the
micro- phone or power adapter, while the 3-pin male connector plugs into the
recorder or mixer. XLR connectors have a locking mechanism on the metal
casing to keep it from coming loose during recording. To insert it, line up
the pins and holes and push on the connector until you feel it snap in place.
To disconnect it, press down on the safety lock while pulling back on the
connector in a steady motion. Never pull directly on the cable wire when
attempting to disconnect it. Doing so can weaken or sever the solder joint
between the wire leads and the terminal posts, thus breaking the circuit and
rendering the cable useless.
FIGURE 11.27
An assortment of
connectors and
adapters used often in
audio production.
RCA CONNECTORS
In the early 1940s, the Radio Corporation of America designed the RCA
phono plug for connecting phonographs (or record players) to amplifiers.
Today, it is used for connecting both the audio and video signal paths of a
diverse array of A/V devices, including television monitors, gaming consoles,
projectors, and numerous other things. RCA plugs are often color-coded to
match the receiving end of an RCA jack. Yellow designates the composite
video channel, while red and white refer respectively to the right and left
audio channels in a stereo system. You’ll also find RCA plugs used for
component video—a video format that uses a separate cable for the red, green,
and blue signals.
ADAPTERS
While it’s best to avoid using adapters, you’ll sometimes need to use one for
hooking up cables and devices with incompatible connectors, and it’s a good
idea to have some good quality adapters in your audio kit. One of the prob-
lems you may run into is the need to mix balanced and unbalanced audio
gear. For example, the external microphone jack on a consumer camcorder
(if it has one at all) is likely going to be a 1⁄8-inch mini plug. In order to use a
profes- sional microphone with a camera of this type, you’ll need an XLR-to-
mini plug
352 SECTION 4 Time-Based Media
adapter. Since the adapter is only traveling a short distance from the end of
the microphone cable to the camera, you shouldn’t have to worry too much
about interference affecting the audio signal.
Still, adapters can complicate matters, as they introduce another potential
failure point in the audio chain. If the adapter dies or comes undone, you’ll
lose your audio. To be safe, secure adapters with a small piece of gaffer tape
to keep them from working lose during the recording. While gaffer tape
looks similar to duct tape, don’t confuse the two. Gaffer tape uses a
different type of adhesive that doesn’t leave a residue when you remove it.
Mistake the two and you’ll end up with messy gear. Make the mistake when
taping cables down on a carpet and expect not to be invited back to the
location.
Tech Talk
FIGURE 11.28
Nobody likes working with a tangled cable. Here, a student uses the over-under wrap to properly coil a microphone
cable.
354 SECTION 4 Time-Based Media
AUDIO MONITORING
One of the simplest recording scenarios is the one-person interview. All
that’s needed is a subject, a recording device (camcorder or audio recorder), a
micro- phone, an XLR cable, and a set of headphones. The producer’s goal is to
acquire source material by interviewing the subject and recording his or her
voice to disk. To achieve professional results, you need to monitor the audio
signal as it is being recorded. Audio monitoring is a two-step process that
includes (1) the objective act of measuring sound intensity and setting the
record levels, and (2)the subjective act of listening to the audio signal as it is
being recorded.
FIGURE 11.29
Like most professional
camcorders, the
Sony HXRNX70U
comes with two XLR
microphone inputs.
An audio control panel
provides user access
to switches and dials
for managing signal
flow and setting record
levels.
Audio Production CHAPTER 11 355
FIGURE 11.30
The analog VU meters shown on top have a range of –20 dB to +5 dB. The digital VU meter
on the Zoom H4n audio recorder (bottom) has a wider range beginning at –48 dB and
ending at 0 dB. Going past 0 dB doesn’t mean clipping will occur, but it does mean you are
getting
precariously close to the distortion threshold. The Zoom recorder features a clipping indicator to
the right of the scale. A red light turns on whenever clipping occurs.
A VU meter displays the strength of the microphone signal (in decibel units)
after it has passed through the preamp (see Figure 11.30). An analog VU meter
has a typical range of –20 dB to +3 dB. A bouncing needle indicates the loudness
of the signal as it modulates throughout the full dynamic range of the record-
ing (from the quietest moments to the loudest ones). Digital VU meters vary in
style. Most of them have a wider range on the low end, starting at –48 or –36 dB.
Instead of a needle, they often feature a row of colored LEDs.
On most VU meters, the region above 0 dB is color-coded red to indicate
exces- sive amplification. If too much amplification is applied, waveform
distortion can occur causing a phenomenon known as clipping. Clipping
permanently corrupts the fidelity of the audio signal and cannot be repaired.
On the opposite end of the scale, you need to also avoid setting the record
level too low. A low audio signal will need to be boosted to acceptable levels
in postproduction. Whenever you re-amplify a recorded audio signal, noise is
introduced, and the quality of the original recording deteriorates. The lower
your original record levels are, the more you will need to re-amplify them
later, and the worse the results will be. Setting levels properly the first time is
the key to getting professional results.
356 SECTION 4 Time-Based Media
HEADPHONES
You should invest in at least one set of good-quality headphones. Look for an
over-the-ear rather than an earbud design. Professional video cameras and
audio recorders will have headphone jacks that let you monitor your audio
during the capture process—use headphones when you are working! Doing so
will help you make sure nothing goes wrong with your audio. You can’t rely
on the VU meter alone to tell you that you are capturing good audio. It
doesn’t tell you that the audio is good or whether or not your microphone is
working or picking up the main subject. You could just be picking up static or
background noise. You might even be recording with the wrong microphone by
mistake, for example, using an internal mic when you meant to be recording
with an external mic. Try to keep the headphones on whenever you are capturing
audio. Just because the audio sounded good when you started doesn’t mean it
will sound good all the way through. Bat- teries die, cables get unplugged,
and microphones can slip. Oh, and cellphones can wreck havoc on your audio
if you are not using RF-shielded microphones. You need to know if your audio
has problems before you go back to the studio to edit.
Listen for potential audio problems—your headphones can help you here as well.
Is your subject wearing metal bracelets that might clink together? What about
background noise? Is the air conditioner making too much noise? What about
that high-pitched electronic squeal from the video projector? Whenever possible,
either eliminate the source of the unwanted sounds or find a better location.
What- ever you do, don’t just say “I’ll fix it in editing.” There’s a good chance
Audio Production CHAPTER 11 357
you won’t be able to, at least not easily, particularly if the noise is in the same
frequency
358 SECTION 4 Time-Based Media
range as your talent’s voice. Watch for echoes in a room and pay attention to
where you set up your equipment. Try to avoid recording in the middle of a large
room with hard walls—instead move to the side of the room. Look for things
that will absorb rather than reflect sound. Recording on a busy street? Your
head- phones will help you make sure you’ve placed your microphone in a
position to minimize the traffic noise.
All too often, newcomers to multimedia production don’t pay enough attention to the
CHAPTER SUMMARY
quality of their audio work and it shows in the final product. To work in the industry, you
need to understand the correct tool to use for a task. For audio, this means
knowing which microphone to use when, and what the relative advantages are for
each type. As an example, we’ve seen that a ribbon microphone is great for using
as a tabletop microphone for conducting an interview in a studio. On the other hand, it
is not a great microphone for conducting an interview outdoors, as it is very susceptible
to wind noise. Make sure you know when to use a balanced cable and when you can
get away with using an unbalanced cable—the short answer is, use balanced
whenever possible, but by all means keep your unbalanced cable runs to around 20
feet or less. Use profes- sional-grade equipment whenever possible—this doesn’t
mean the most expensive, just good quality—and avoid internal microphones when
possible. Oh, and don’t forget your headphones. They really are one of the most
important tools you’ve got, not only when you are editing audio but when you are
capturing it as well. Few things are worse in audio production than coming back from an
assignment and realizing that the battery died in your microphone three minutes into the
interview and, because you weren’t wear- ing headphones, you didn’t realize it.
Audio Production CHAPTER 11 359
353
Prior to 1956, motion picture film was the only recording medium available
for capturing and storing televised images. Using a device called a kinescope,
a 16mm or 35mm motion picture camera was set up to record electronic
images
Multimedia Foundations. DOI: 10.1016/B978-0-240-81394-3.00012-4
© 2012 Elsevier Inc. All rights reserved.
354 SECTION 4 Time-Based Media
FIGURE 12.1
The popular children’s program Howdy Doody was telecast by NBC from 1947 to 1960. Prior
to the invention of video recording, programs like this one were produced and transmitted “live”
to the home television audience.
directly from the surface of a television monitor (see Figure 12.2). This pioneer-
ing method of video transcoding was used to generate a photographic archive of
scanned television images in real-time. Transcoding is the process of converting
media from one form, or format, into another. Thus, in the early days of
broad- casting the only way to make a recording of a television signal was to
trans- code it from its native form, as a scanned projection on the surface of a
cathode ray tube, to an optical reproduction of light on photographic film.
While the filmed image represented aspects of the original form, transcoding
resulted in the creation of something entirely different as well. For example,
while a film could be shown to a local audience gathered in a theater using
a mechanical
Sound and Video Recording CHAPTER 12 355
FIGURE 12.2
For many years, the kinescope was the only system available for recording a “live” television
production.
one must always be mindful of the negative consequences that can occur
when- ever television signals or video files are converted from one format into
another. Information is always lost during the process of transcoding.
MAGNETIC RECORDING
Ampex Corporation invented the world’s first open-reel tape recorder in 1948.
Dubbed the Model 200, it was used by the American Broadcasting Company
(ABC) for the tape-delayed transmission of The Bing Crosby Show. The event
went down in history as the first ever tape-delayed broadcast of a radio program
in the United States. It took Ampex another eight years to release the first
commercial videotape recorder, the VRX-1000 (or Mark IV), which they
introduced at the National Asso- ciation of Radio and Television Broadcaster’s
convention in Chicago in 1956.
With an audio recorder, the magnetic tape moves continuously across the
path of stationary record and playback heads. Only the tape is in motion (see
Figure 12.3). On a system like this, the audio signal is recorded
longitudinally, across the length of a tape, as an uninterrupted stream of
invisible data. Video- tape recording is much more complex. Because of the
large bandwidth require- ments of television, video signals are written to tape
using multiple record heads.
FIGURE 12.3
With an open-reel (or
reel-to-reel) recorder,
an analog audio signal
is recorded to the
surface of ¼-inch
wide magnetic tape.
The signal is captured
on tape by the record
head and played back
a split second later
by the playback head.
During recording and
playback, the tape is in
constant motion. The
heads, however, are
358 SECTION 4 Time-Based Media
stationary.
Sound and Video Recording CHAPTER 12 359
For example, the VRX-1000 was called a quadruplex (or quad) recorder because
it used four record heads for writing data to a magnetic tape that was 2 inches
wide. The record heads were mounted 90-degrees apart along the perimeter
of a headwheel assembly that would spin at a high rate of speed as the tape
(also moving) traversed in the opposite direction across its path. The
invention of the videotape recorder was such an important breakthrough for
the broadcasting industry that in 1957, the Academy of Television Arts and
Sciences recognized Ampex Corporation with an Emmy award for its
accomplishment.
GREAT IDEAS
Helical Scanning
With an audio tape recorder, the signal is recorded to tape as it travels past a stationary record head at a constant rate of sp
bandwidth video recording. In 1961, Ampex invented helical scanning, a technology that has been used in tape-based camcor
FIGURE 12.4
An inside view of a video cassette recorder (VCR) reveals the slanted headwheel assembly used in a helical scanning recor
(Continued)
360 SECTION 4 Time-Based Media
a portion of several video frames. For this reason, unless you are performing a repair,
and you know what you are doing, you should never attempt to edit videotape by
hand.
Unlike motion picture film, the frame and track regions on a magnetic tape are
invisible to the human eye. But just because you can’t see them doesn’t mean they
don’t exist (see Figure 12.5). For recording purposes, videotape is divided into four
tracks. The video track is the widest region and is used to store the picture portion
of a television program. Videotape also contains two audio tracks. The audio tracks
can be used in tandem for recording a two-channel stereo signal or separately for
recording two monaural feeds. For example, in a two-person interview, the microphone
signal from Person A can be recorded to audio track one, while the signal from
Person B is routed to audio track two.
The fourth region is designated as the control track. To understand the purpose of
the control track, it is helpful to draw a comparison once again to motion picture film.
Film contains two lines of sprocket holes lined up along each edge of the filmstrip.
As a film is projected, moving gear like wheels engage the sprocket holes to
advance one frame of film to the next. The sprocket holes keep the film aligned and
moving at the appropriate speed. Like film, videotape is in constant motion during
record and playback modes. As the video and audio signals are recorded to tape, an
electronic synch pulse is written to the control track to mark the location of each
video frame. NTSC video systems generate 30 frames (or still images) of picture
information every second. The control track marks where these frames begin and
end in order to ensure that the videotape plays correctly.
FIGURE 12.5
The surface area of a videotape is divided into linear regions called tracks. Most tape-
based systems support two tracks for stereo audio, one track for video, and one track for
synchronization or control data.
Sound and Video Recording CHAPTER 12 361
Tech Talk
The 2-inch quad format continued to evolve throughout the 1960s and 70s as fea-
tures like color recording, helical scanning, and portability were incorporated
into the design of newer, more advanced models. One of the biggest changes in
open-reel designs occurred in 1976 when Ampex and Sony introduced the Type
C videotape recorder (VTR). Type C VTRs used a narrower tape that was one-
inch wide, cutting the physical size of the media and storage reels in half, while
keeping the recording time the same (1 hour per reel). Sony also designed a
portable model that was used for many years in remote field production. Type C
VTRs included popular features such as stable freeze frames, variable playback
speeds (slow-mo and fast motion effects), and advanced jog and shuttle control.
The one-inch videotape format was the dominant recording medium for
broadcast television for nearly two decades.
FIGURE 12.7
A producer conducts an interview as her camera operator records the subject. Prior to 1971,
field recording like this was done entirely with film cameras. This meant a time-delay, because
film had to be processed before it could be viewed, edited, and aired. Today, electronic images
and sound can be shot, viewed, and transmitted instantaneously using either a tape-based or
tapeless recording system like the one shown here.
Bruce C. Murray/Shutterstock.com
U-MATIC ¾-Inch
In 1971, the Japanese consortium of Sony, JVC, and Matsushita (Panasonic)
led the way in pioneering a new era of electronic newsgathering (ENG) with
their joint release of the U-matic videotape-recording format. U-matic, better
known as ¾-inch because of its smaller tape width, offered field producers
the first viable alternative to a film-based production workflow. The most
appealing feature of the ¾-inch format was that the tape was enclosed in a
portable plastic cassette. When inserted into a U-matic VCR (Video Cassette
Recorder), the tape would be threaded automatically. The cassette housing
provided protection for the tape and eliminated the process of threading by
hand.
Compared to open-reel recording systems, ¾-inch decks were cheap and had
a relatively small form factor. The portable version of the U-matic line was
light enough to be carried by one person using a shoulder strap. By
supporting the unit with the neck or shoulder, a tape operator could maintain
convenient access to the controls while roaming freely during a mobile
recording session. Before the invention of the camcorder, the video camera and
recorder were two separate devices. A mobile production crew required at
least two people—one person to
Sound and Video Recording CHAPTER 12 365
FIGURE 12.8
VHS, Betacam, and Hi-8 are analog tape formats. MiniDV is a digital tape format.
operate the camera and compose shots, and the other to monitor the video
and audio signals during recording. The camera and VCR were tethered by a
cable, which meant that the tape operator had to stay close to the shooter at
all times. U-matic field recorders could be run on AC power or with a
rechargeable Ni-Cad battery, further enhancing their usefulness in remote
locations. The portable units used a small cassette with enough tape for 20
minutes of uninterrupted recording.
U-MATIC SP
U-matic is regarded as one of the most successful and longest lasting video
for- mats of all time. In 1986, Sony released U-matic SP (for Superior
Performance), which used chrome tape and enhanced signal processing to
produce a better picture. It remained, however, a composite videotape
format.
Tech Talk
FIGURE 12.9
Like most professional nonlinear editing programs (NLEs), Avid Media Composer includes a color corrector interface for
adjusting the color properties of a video clip. Color grading is the process of enhancing the color presentation of a film or
video project during the final stages of the postproduction process.
368 SECTION 4 Time-Based Media
fared better than its predecessor (the failed M format), Betacam SP became
the dominant professional tape format throughout the 1990s.
D1 AND D2
In 1986, Sony and BTS collaborated in bringing the first digital tape format
to market. Dubbed D1, the format provided a way for broadcasters to
digitally write an uncompressed component NTSC video signal to tape.
Unfortunately, D1 tape and the equipment required to use it was expensive
and terribly inef- ficient in terms of managing bandwidth (1 MB of storage was
required to write a single uncompressed frame of video). D1 had limited
success, ultimately being adopted by only a relatively small number of high-
end users who had deep pockets, and who were most interested in using it for
motion graphics, compos- iting, and special effects. In 1988, Ampex and Sony
released D2, the first digital tape format to be widely accepted by professional
users. In their second attempt to design a cost-effective digital tape format,
engineers succeeded by fixing the bandwidth issues that had plagued D1.
They accomplished this by switching back to composite video sampling. D1
and D2 were cassette formats that used
¾-inch-wide tape.
FIGURE 12.10
Tape formats are generally not interchangeable. For example, while the professional VTR on the left can record and play
back HDCAM and Digital Betacam cassettes, it cannot be used to play a VHS tape. For that format, you will need a VHS
unit like this vintage model pictured right.
Sound and Video Recording CHAPTER 12 369
PANASONIC D3
Panasonic released the D3 format in 1991. D3 decks were able to write
uncom- pressed composite video onto a ½-inch wide tape, the narrowest yet
for a professional digital format. Given its smaller form factor, Panasonic’s
D3 tape stock and recording equipment were considerably less expensive to
buy and operate than D2. The ½-inch-tape width also meant smaller cassettes
and longer record times.
FIGURE 12.11
Large-format camcorders have a larger form factor and are often easier to use for shoulder-
mounted handheld camera work. Small-format cameras are less expensive, and until recent
years have generally lacked the image quality and features that professionals need. However,
the gap between high-end professional and low-end professional is shrinking as camcorder
technology and recording formats continue to improve.
370 SECTION 4 Time-Based Media
MPEG IMX
Marking a potential end to their long run of success in the professional video-
tape market, Sony released MPEG IMX in 2001. This advanced digital HD
format is fully compatible with the DTV television standard. It is also provides
full sup- port for playing back all of the legacy Betacam formats (Betacam,
Betacam SP, Digital Betacam, and Beta SX).
DV Formats
DVCPRO AND DVCAM
Sony and Panasonic developed competing professional variants of the DV for-
FIGURE 12.12
High-Definition mat. Panasonic introduced DVCPRO in 1995. The format was marketed to low-
signals require end professional users (local ENG news crews, etc.) looking for an alternative to
more bandwidth for expensive digital formats like Digital Betacam. Panasonic achieved professional
recording. The old results with DVCPRO by switching to metal particle tape, widening the track
formats would once
pitch from 10 microns to 18, and increasing the speed of the tape transport by
again be surpassed by
new ones, with names 80%. While the faster tape speed improved record quality, it cut record time on
such as HDCAM, a mini-cassette from 60 minutes to 12. To compensate, Panasonic introduced a
DVCPRO HD, and HDV.
Sound and Video Recording CHAPTER 12 371
FLASHBACK
DV: The Blue Book Standard
In 1995, a large consortium of ten consumer electronics companies, including long-
time rivals Panasonic (Matsushita) and Sony, released DV, an international set of
standards for digitally compressing a component NTSC (720486) or PAL
(720576) video signal to tape. DV (short for Digital Video), was more than just a
tape format,
it was a recipe book that manufactures could use for designing a variety of recording
formats for different market segments. Known in engineering circles as Blue Book
or IEC 61834, DV was quickly adopted by the consumer electronics industry
and prosumer marketplace (see Figure 12.13).
DV uses intraframe compression to reduce spatial redundancy on a frame-by-frame
basis. DV audio is recorded on one of two tracks as an uncompressed 16-bit Linear
PCM stream with a sample rate of 48 kHz. The sample rate is slightly higher than
the
44.1 kHz rate used in audio CDs. As a result, most NLEs have adopted 48 kHz as the
native sample rate for the audio portion of a DV recording. When DV footage is ingested
into a computer, it is saved either as raw data, or wrapped in a container format such as
Audio Video Interleave (AVI), QuickTime (MOV), or Material eXchange Format (MFX). The
NLE system determines which wrapper to use. For example, Final Cut Pro wraps DV in a
QuickTime container, Apple’s preferred video format; Avid Media Composer uses MXF.
While designed primarily for recording to digital tape, DV streams can also be written
directly to optical disk, flash memory cards, and computer hard drives. It is a flexible
hybrid format that helped to prepare the way for an industry transition from tape-based
to file-based recording.
The Blue Book standard spawned a flurry of activity in both the consumer electronics
and pro video markets as manufacturers sought to profit from the world’s first digital
tape standard with broad mass-market appeal.
FIGURE 12.13
The Canon XL-1 was a prosumer standard-definition camcorder made from 1997 to 2001.
It used DV tape and was popular with corporate producers and independent filmmakers.
372 SECTION 4 Time-Based Media
HDV
As the inevitable transition from SD to HD loomed on the horizon, a group
of manufactures looked for a way to record high-definition television signals
on standard DV tape. The answer came in 2003 when a consortium of four
com- panies (Canon, JVC, Sharp, and Sony) announced the HDV standard.
In order to record an HD signal within the same footprint as DV, engineers
switched to an interframe compression scheme. They also decreased the
resolution of the HD image to 14401080. Full HD has a native resolution of
19201080. HDV offered an affordable way for consumers, prosumers, and
low-end profession- als to work in HD without having to migrate to a full
bandwidth HD workflow. And while HDV has been used, on occasion, in the
production of professional broadcast programs, it is best to think of it as an
advanced consumer or prosumer format. Most HDV devices are backward
compatible, allowing users to record in standard DV as well.
DVCPRO HD
The final addition to the DV family of tape formats is DVCPRO HD, a low-
compression format with features similar to Sony’s HDCAM, such as 4:2:2 color
sampling, a high bit rate, and intraframe compression. DVCPRO HD is used
almost entirely by professionals.
same color value across the entire sequence of 150 frames. This phenom-
enon is called temporal redundancy and occurs whenever the value of a pixel
remains unchanged from one frame to the next in a time-based sequence.
And while the identical blue pixels in our video example extend forward over
time, they also stretch outward within the space of each individual frame.
This phenomenon is called spatial redundancy (see Figure 12.14).
The goal of video compression is to conserve bandwidth by eliminating
spatial and/or temporal redundancies during the recording process (see
Figures 12.15 and 12.16). Two methods of video compression are most
commonly used today.
The first method is called intraframe (or I-frame) compression. This approach elim-
inates spatial redundancies “within” a video frame in much the same way
that JPEG compression is used to reduce them in a digital photograph. I-
frames are typically compressed at a ratio of 10:1. This means that a
compressed I-frame consumes as little as 10% of the file space of a raw
uncompressed frame. Since I-frames are fully defined by information from
within the frame, they can be easily decoded and rendered onscreen during
playback and editing. However, the overall amount of compression that’s
achieved with this approach is limited since temporal redundancies are not
addressed.
The second and more common method is called interframe compression, a tech-
nique that exploits both spatial and temporal redundancies. Using the previ-
ous method of intraframe compression, all of the frames in a video stream are
turned into I-frames. Each one is intracoded to eliminate spatial redundancy.
Thus, compression is applied evenly to all of the frames within a video
stream. With interframe compression, I-frames are created at fixed intervals
(typically every 15 frames). An I-frame marks the beginning of a packaged
sequence of adjacent frames called a GOP (Group of Pictures). The fully
defined I-frame serves as a keyframe or reference for other frames in the
group. Its job is to hold repeating color values in place that will not change
across the sequence. Basi- cally, it creates a digital marker on the frame that
says “do not change the color of this pixel until you are told to do so.”
HDV is a popular high-definition video format that uses an interframe com-
pression algorithm called Long-GOP MPEG-2. With Long-GOP, each I-frame is
followed by a sequence of 14 frames designated as either a P-frame or a B-
frame. A P-frame is a predictive coded image that only stores data for pixels that
are dif- ferent from the preceding frame. For example, in a shot of a bird
flying across a blue sky, only pixels related to the moving bird would be
encoded to a P-frame. The unchanged background pixels simply carry forward
from information stored in the previous frame. On average, a P-frame can be
compressed twice as much as an I-frame. A B-frame is a bidirectional
predictive coded image. It records changes from both the preceding frame
and the one immediately following. On average, a B-frame can be
compressed twice as much as a P-frame. With interframe encoding, all
frames are not equally compressed. The more heavily compressed P-frames
and B-frames require more processing power to encode and decode.
374 SECTION 4 Time-Based Media
FIGURE 12.14
Spatial redundancy (top) is that which occurs within a single still image or video frame. Temporal redundancy (bottom)
occurs over time across a sequence of motion picture frames. Video compression deals with both kinds of redundancy.
Sound and Video Recording CHAPTER 12 375
FIGURE 12.15
Apple Compressor 4 is a professional video compression and encoding tool. You can use this program or others like it
to encode your video in any number of common distribution formats. For example, you may want to produce a DVD or
Blue-Ray disc, upload the video online to YouTube or Vimeo, or save it to your iPad or Android tablet. Your video must
be encoded in a format that is appropriate to each distribution channel you plan to use. For this reason, editors often
find themselves creating multiple versions of a single project for different users.
FIGURE 12.16
While Compressor 4
is a powerful encoder,
it is a commercial
product that will only
run on a Mac. MPEG
Streamclip (pictured
here) is a free
crossplatform video
encoder. You can use it
to convert digital video
files to many of the
common formats used
today.
376 SECTION 4 Time-Based Media
Depending on the power of the computer and the software being used, editing
systems can experience difficulty when playing back interframe streams on
the fly. Remember, only the I-frames are fully coded with all of the color
informa- tion needed to render it on screen. In order to properly render P-
frames and B-frames, the playback engine needs to access relevant picture data
that’s stored in other frames in the GOP sequence.
CONSUMER FORMATS
For the longest time, consumers had no ability to record television shows off
the air in their homes. If you missed a show, the only option was to wait for
the station to re-air it. A rerun might take several months or never happen at
all. Likewise, if you missed the theatrical release for a motion picture film,
chances were it would never be shown publically again. The likelihood of it
appearing on television was just as remote. Home video recording formats
would change this forever, giving consumers the ability to record a favorite
program for viewing at a later time. This new consumer behavior would become
known as time-shifting. The invention of the home VCR would also spawn the
movie rental industry, giving Hollywood a profitable revenue source for the
redistribution of motion picture films to audiences in the comfort of their
own living rooms.
Tech Talk
(Video Home System). For several years both formats competed side-by-side
in a divided marketplace. Consumers were often confused. Video stores were
required to stock two versions of a film. And film distributers had to deal
with the hassle and expense of supporting two incompatible formats. Despite
the fact that Betamax was first to market and that it produced a sharper
picture, the marketplace eventually went with VHS. The decisive factor came
down to record time. VHS offered two hours of record time in standard play
(SP) mode. This was double what Betamax could deliver. In the end, VHS was
a better fit for the movie industry. For twenty years, VHS reigned as the home
recording format of choice throughout most of the world. VHS-C and Super
VHS-C were compact versions of the larger cassette formats, designed for use
in VHS camcorders.
MiniDV
DV was originally released to the public under the name of DVC (for Digital
Video Cassette). DVC uses ¼-inch wide metal evaporated tape, which is capable
of recording a high-quality video signal. The cassettes come in two sizes.
The large cassette, designed primarily for desktop recorders, can hold three
hours of recorded footage. The small cassette, called MiniDV, has a maximum
record time of one hour and is intended primarily for use in consumer
camcorders. How- ever, because of its popularity as a flexible multiuser format,
MiniDV tapes come in many different flavors ranging from consumer-grade to
master-grade quality.
GREAT IDEAS
Optical Disc Formats
In 1995, Panasonic, Phillips, Sony, and Toshiba jointly released the DVD-Video standard.
Identical in look to an audio compact disc, DVD was designed as a distribution format to replace Betamax and VHS (see ). DV
The original DVD format uses a red laser beam to record and play back an MPEG-2 encoded video stream. The pitch of the re
4.7 GB on a standard single-sided disc. While this is enough bandwidth for storing a standard-definition movie, it is far below
In order to increase the storage capacity of optical discs, engineers looked to the opposite end of the color spectrum. At 405
(Continued)
378 SECTION 4 Time-Based Media
released the Blu-ray format, a high-density optical disc for distributing movies in HD. A Blu-ray disc is t
FIGURE 12.17
DVD replaced VHS as the main distribution format for the motion picture industry. The market shifted a
Tech Talk
MP3
MPEG’s first project was published in 1993 under
the name of MPEG-1, which is most remembered
for establishing the popular MP3 audio format.
MP3 was the first digital codec designed for stor-
ing music on a portable audio player. MP3 offered
users the ability to compress CD-quality audio (44.1
kHz) by a ratio of 10:1. In addition, MPEG-1
video was established, along with the first set of
specifi- cations for encoding SD video using
intraframe or interframe encoding. Unfortunately,
MPEG-1 video had a low bit rate of 1.5 Mbps,
which was similar in quality to VHS.
MPEG-2
MPEG-2 was approved in 1993 as an advanced stan-
dard for encoding SD and HD video formats at a
significantly higher bit rate. It would become the
international standard for encoding SD video on a
red laser DVD. Audio compression was enhanced
as well. With MPEG-2, came Advanced Audio Cod-
ing (AAC), a format that extended the stereo capa-
bilities in MPEG-1 to multichannel surround
sound (Dolby Digital 5.1). In addition to its use as
the encoding standard for DVD-Video, MPEG-2
has been incorporated into numerous other
formats like HDV, XDCAM, consumer formats
like MOD and TOD, and the terrestrial HDTV
broadcasting standards ATSC and DVB. MPEG-2 is
a fluid stan- dard that has been amended several
times since its initial public release.
FIGURE 12.18
The sequence preset window in Adobe Premier CS 5.5 lists a
slew of video formats supported by the NLE. While a list like
this can be daunting at first to understand, most of the time
you will find yourself working with a handful of trusted video
formats and codecs. So don’t worry, you don’t need to know
them all.
Sound and Video Recording CHAPTER 12 381
Tech Talk
The actual codec goes by many different names including AVC, H.264, MPEG-
4, and H.264/MPEG-4. AVC is best known for being one of several native
codecs supported by Blu-ray Disc. Both ATSC and DVB have approved AVC
for use in terrestrial HDTV broadcasting. AVC is currently one of the most
popular formats used for uploading video to online hosting sites like YouTube
and Vimeo.
FIGURE 12.19
In lieu of videotape,
the Sony XDCAM EX
series of professional
camcorders records
video to an SxS
memory card.
P2
P2 is a solid-state memory card format introduced by Panasonic in 2004.
These memory cards are capable of storing video streams in any of the follow-
ing formats: DV, DVCPRO, DVCPRO25, DVCPRO50, DVCPRO-HD, or
AVC-
Intra. AVC-Intra is a master-quality recording format developed by Panasonic
for its top-of-the line broadcast cameras. A P2 card contains multiple SD cards
packaged together inside a large-format metal case known as a PC or PCMCIA
Card.
AVCHD
One of the goals of the MPEG-4 initiative was to specify an open standard for
recording HD video that would also be compatible with inexpensive flash mem-
ory card formats such as Secure Digital (SD and SDHC) and Compact Flash
(CF).
Sound and Video Recording CHAPTER 12 383
FIGURE 12.20
Proprietary card formats like SxS and P2 are designed for professional users. These cards are
significantly better than off-the-shelf SD memory cards, and as a result are more robust and
reliable. Because of this, they are also more expensive, a fact that puts them out of reach for
many low-budget producers.
This goal was realized in 2003 with the development of H.264/MPEG-4 AVC, a
high-quality compression and recording standard that has become widely popu-
lar in recent years.
Tech Talk
1
Sony Corporation. (2009). AVCHD Introduction Handbook (2nd Edition).
384 SECTION 4 Time-Based Media
A Comparison of Record Time (minutes), Record Format, and Memory Card Capacity
Table 12.2
FIGURE 12.21
The AVCHD standard
specifies this file
directory structure for
recording video and
metadata to a solid-
state memory card. It’s
important to keep this
file structure intact.
If any of the folder
names is altered,
your editing software
may not be able to
recognize the video
files. With AVCHD,
video files are saved
to the STREAM
folder in the .MTS
format (MPEG
Transport Stream).
In 2006, Sony and Panasonic jointly released a variation of the MPEG-4 stan-
dard called AVCHD. Originally designed for consumers, AVCHD is a
proprietary format that has been incorporated into a new fleet of prosumer and
professional camcorders. This format has grown quickly in popularity and has
been licensed for use by other manufacturers, including Canon, JVC, and
Hitachi. The AVCHD recording format is used in Sony’s NXCAM line of
cameras as well as Panasonic’s AVCCAM series. A consumer-grade version of
the standard, called AVCHD Lite, has a maximum resolution of 720p and is
often used in digital still cameras for shooting HD video (see Figure 12.21).
Video recording began in 1956. Since that time, a host of formats have come and
CHAPTER SUMMARY
gone as engineers and product developers have found newer and better ways of
encoding video and audio signals. For many years, the recording device (the camera or
VCR) was tied to a proprietary medium such as a magnetic tape reel or videotape
cassette. A video recording could only be played back if you had the right hardware for
viewing it.
The most significant development in recent years is the development of solid-
state technologies that use file-based recording formats and codecs. Such formats
have dras- tically driven down the cost of professional HD recording systems and
media and are often more flexible and convenient to use. However, because they
are more flexible, there is hardly ever just one way to record or transcode a video or
audio signal. You can’t
386 SECTION 4 Time-Based Media
just turn the camera on, press the record button, and expect the best results. You
need to know a lot more information related to recording formats, codecs,
container files, resolutions, frame rates, sampling rates, bitrates, etc. in order to make
informed judg- ments about the best resolution and quality for a given job or
assignment. The more you know now, the better positioned you will be when everything
changes once again—and the next generation of recording technologies emerges.
CHAPTER 13
Time-Based Editing
383
NONLINEAR EDITING
Video editing is the art of arranging static and time-based media assets into a
linear form of presentation for the purpose of telling a story or
communicating a message. The goal of editing is usually to produce a
thoughtful narrative with a clear beginning, middle, and end. Audio editing is
similar but with a focus on sound-based media elements only. Many of the
basic editing techniques used in video are applicable to audio, particularly if
you are using professional audio editing software.
Motion picture editing began as a simple manual process. In the early days of
film, an editor would review and edit footage using a hand-operated device (see
Figure 13.1). Editing was performed by physically cutting film into sections,
rear- ranging the pieces into a new linear order, then splicing them together
again using clear tape or another type of adhesive—hence the term linear editing.
This method was also used in audio tape editing. You’ll find that many of the
terms used in digi- tal audio and video editing today have their origins from this
era, particularly the use of terms such as the razor blade tool referring to the old
cutting implements, and
FIGURE 13.1
Motion picture editing used to be a physical act that required manually viewing, cutting, and
splicing pieces of film together to make a movie. Today, film and video editing is done virtually
using an NLE program. There’s no more mess and no more film left behind on the cutting room
floor.
the term bins, referring to the actual bins that film editors used to hold segments of
film and audio tape.
While film editing began as a physical process, video editing started out as
an electronic process. Unlike film, one cannot expose videotape to the light
in an effort to view the semitranslucent frames of each still image. Video
images are processed electronically, which means that viewing and editing
has to be done using electromechanical players, recorders, and monitors. In
the beginning, video editing was nothing more than systematic duplication.
Material was copied section by section from a recorded tape in one VCR to
a new tape in a second VCR, a process known as tape-to-tape or machine-to-
tape editing. This technique was also used in audio production, though it was
supplemented by multitrack recorders, which allowed multiple audio tracks
to be recorded and edited on a single piece of audiotape.
Machine-to-machine editing worked pretty well for many years. The simplest
machine-to-machine editing system involved five pieces of equipment: (1) a
playback deck; (2) a playback monitor; (3) a record deck; (4) a record
monitor; and (5) an edit controller. The edit controller was connected to both
decks, allow- ing the editor to control them remotely. To perform an edit, the
editor would begin by “electronically” marking the beginning and end of a
selected shot on the source deck. This involved shuttling the deck (moving or
scrubbing the tape backwards or forwards), pausing it at a specific spot on the
tape, and pressing physical buttons to mark the respective IN and OUT
points. The editor would perform the same procedure to set an IN point on
the record deck. Pushing one last button allowed the editor to preview or
perform the edit. During the pre- roll phase, both machines would back up 3–
5 seconds ahead of the designated
Time-Based Editing CHAPTER 13 385
IN points and then would roll (start to play) in unison. When the IN point
was reached on the record VCR, recording would begin and continue until
the OUT point triggered the end of the edit.
This type of machine-to-machine editing configuration was sometimes called a
cuts-only system, because the only type of edit you could perform was a cut. If
you wanted to create anything other than a straight cut transition, you had to
use an A/B Roll editing system. The A/B Roll workstation had two source decks
instead of one. The A-Deck was loaded with the primary footage (interviews,
scripted dia- log, etc.) while the B-Deck contained secondary support material
(cover shots, establishing shots, cutaways, etc.) that would appear while the
main person or character was talking on screen. This is where the familiar,
yet sometimes con- fusing, term B-Roll comes from.
The computerization of media has made machine-to-machine editing a thing
of the past. The edit controller is no longer a physical device sitting on a
tabletop, but rather software powered by a computer. Welcome to the age of
nonlinear editing (NLE), where programs such as Adobe Premiere, Avid Media
Composer, and Apple Final Cut Pro, are used by professional editors to craft
their stories. For consumers, the marketplace offers a number of simple-to-
learn NLEs like Pinnacle Studio, Apple iMovie, and Windows MovieMaker.
Professional NLE titles are designed for filmmakers and producers who want
advanced tools and need to maximize their workflows. Audio editing
software comes in a similar range, from consumer-level software such as
GarageBand to professional-grade software such as Logic, Audition, or
ProTools.
FIGURE 13.2
Television news anchors are able to maintain eye contact with the viewer by reading their
scripts via a teleprompter mounted to the front of the camera.
scene off-book, or without the aid of a written prompt or script. This means
that the talent will have to memorize lines and rehearse them prior to the
actual recording. On-camera performances are rarely perfect, and directors
often ask talent to make multiple attempts (takes) of a scene before he or she
is satisfied with the performance.
In some situations, dialog may be performed with the aid of a teleprompter (see
Figure 13.2). A teleprompter projects words onto the surface of a glass panel
that’s mounted in front of a television camera. The performer reads the text as it
scrolls upwards and can still talk straight to the camera. Teleprompters are
often used in studio productions (news, sports, talk programs, etc.), but can
also be used in remote field production. While teleprompters can cut down
on rehearsal time and speed production, their use doesn’t guarantee perfect
takes. Reading from a teleprompter is tricky. Some people are better at it than
others.
SOUND BITES
A sound bite is a short segment of a longer interview or speech that is inserted
into a video program or news package. In television, a sound bite is denoted
by the scriptwriting abbreviation SOT (pronounce as S-O-T) for sound on tape.
In radio broadcasting, a short interview segment is called an actuality.
Whereas inter- views are shot or recorded during the production phase of a
388 Section 4 Time-Based Media
project, sound bites
Time-Based Editing CHAPTER 13 389
GREAT IDEAS
The Stand-Up
In television, a reporter
stand-up or on-camera host
segment can be used as an
alternative to a voiceover or
in combination with it. For
example, a reporter may
choose to use a stand-up
at the beginning and end of
a story and voiceovers
throughout the rest of the
package. The best stand-
ups add meaningful visual FIGURE 13.3
value to the story by A TV reporter prepares to deliver an on-camera stand-up.
providing something other than a good look at the host. The stand-up can be used to
demonstrate a key point or guide the viewer through a complex process, where video is
not available or difficult to acquire.
390 Section 4 Time-Based Media
VOICEOVER (VO)
The voiceover (or VO) is a narrative device used to describe an audio segment
that’s recorded off-camera without synchronized video. It is the sound of the
hidden announcer, storyteller, reporter, narrator, or host, who guides the
audience through the program—filling in details while adding color and
continuity to the linear presentation of information. Voiceovers are often used in
conjunction with sound bites to form longer strings of information. In news
reporting, the VO-SOT (pronounced VO¯ -sät) is a technique in which the
television anchor reads a scripted story off-camera, while related images and
sounds appear on screen. VO-SOT-VO describes a fundamental editing technique
where voiceovers and sound bites alternate to form the basic narrative structure
of a linear program.
B-ROLL
B-Roll is video source material that’s used to fill holes in the timeline where syn-
chronized video is not included, or as a source of supplemental footage to
accom- pany a video sound bite. For example, when a voiceover is added to a
sequence, B-Roll is used to supply the pictures. Likewise, in an interview
segment, B-Roll can be used, in part or in whole, to replace the shot of the
person speaking, effectively converting an SOT into a VO. B-Roll includes
things like cut-ins and cutaways, which we will discuss later, and other footage
that visually supports what is being said on screen. For example, as a reporter
is speaking off-camera about a local bank robbery, the viewer sees a
sequence of B-Roll shots depicting the scene of the crime. This could include
material obtained from surveillance cameras along with other shots taken in the
aftermath. There’s a popular saying among editors that you can never get
enough B-Roll. The more B-Roll an editor has to work with, the more options he
or she will have for improving a story’s pace and structure.
NATURAL SOUND
Natural sound (nat sound), or ambience, occurs in the background as a scene
is being recorded. As humans, we are able to distinguish between foreground
and background sounds. The sound of two voices, joined together in
conversation, is registered as foreground sound. The sound of a car driving by is
a background sound. When a camera operator records B-Roll, it is common
practice to record audio that is simultaneously occurring in the background.
You can use nat sound in both audio and video productions to help establish
a sense of place.
TITLES
A title (or super) is a text-based graphic that is superimposed over video to
identify visual content within the frame. One of the most common types of titles
you see on nearly every television show is called a lower third—so named
because of its usual placement near the bottom of the video frame. A lower
third is used most often to identify the name and title of the person speaking
on screen.
A title graphic is placed on screen through a process called keying. Keying
Time-Based Editing CHAPTER 13 391
replaces the background pixels in a title graphic (known as the alpha channel)
with video from another source. This technique is similar to the superimposi-
tion of a weather map behind the weathercaster during the evening news. The
392 Section 4 Time-Based Media
FIGURE 13.4
Adobe Photoshop includes several film and video presets. For example, if you need to create a full screen graphic for an
HD television program, you can choose the one shown here. Film and video presets include visual guides for the action
and titles safe areas.
Full-Screen Graphics
A full-screen graphic takes up the entire screen space and therefore does not
have to be superimposed with another video source (see Figure 13.6). A matte
(usually
Time-Based Editing CHAPTER 13 393
FIGURE 13.5
A lower-third title graphic is superimposed over the video of a television reporter. This one was
generated using a template in Apple Final Cut Pro X.
FIGURE 13.6
A full-screen graphic like this one fills the entire video frame.
GREAT IDEAS
Designing Graphics for Television
When designing television graphics, keep in mind that viewers may be watching the
video on a low-resolution medium, or even in a small window on a computer monitor.
Viewers have a short time frame in which to process information as it is flashes on
the screen. Follow these tips when planning your graphics:
Avoid Clutter. Remember, less is more. Don’t pack more information into the
visual screen space than is absolutely necessary. Avoid long sentences and
paragraphs. Instead, use short phrases and bullet points whenever possible.
Use Thick Sans-Serif Fonts. Select high-density fonts with a thick stroke. Scripts
and other light stroke fonts don’t translate well to video and lose much of their
detail, making them hard to read and frustrating the viewer.
Use Large Font Sizes. Use font size to denote a visual hierarchy of information,
but keep the sizes relatively large. It’s better to use a readable font and break up
information across several screens than to cram too much information into a
single title screen or graphic.
Strive for Good Contrast. If the background is dark, the foreground text should be
light. Separate the text from the background by using drop shadows, stroke, light, and
texture. Text should pop off the surface of the background. Try using a high-contrast
background matte behind text instead of superimposing the text directly over video and
running the risk of making the text hard to read.
Use Video-Friendly Colors. White and yellow are popular colors for screen text
because they are easy to read against a dark background. Black also works quite
well when placed on white or a light-colored background.
FIGURE 13.7
Your NLE software should include a character generator for composing on-screen titles
like the one shown here in Apple Final Cut Pro X. This interface allows you to control the
placement and appearance of the text within the frame.
(Continued)
Time-Based Editing CHAPTER 13 395
Margins and White Space. Stay within the title safe area of the frame—the area where titles will display w
these terms is that older television sets didn’t actually show the whole picture; a mask usually covered
CONTINUITY EDITING
The editing together of shots in sequence should be done in such a way that
the linear presentation of visual and aural elements in a timeline is perceived
as natural and coherent to the viewer. Bad edits can be disruptive to the
narra- tive flow of a story, causing viewers to disconnect, get confused, lose
their place, or stop watching altogether. There’s a saying that “the best edits
are the ones you never see.” In other words, good edits are so seamless and
logical, that the viewer doesn’t even notice when they occur. And this is what
editors often hope to achieve. Continuity editing is a term used to describe a wide
range of practices and techniques that editors use to achieve smooth and
natural transitions from shot to shot over time. When edits are performed
well, the narrative continuity of the story is enhanced—leading to a more
satisfying and enriching experience for the viewer (see Figure 13.8).
Cut-Ins
One of the techniques used in continuity editing is the cut-in. A cut-in directs the
viewer’s attention to a related object or alternative viewpoint within the
current
396 Section 4 Time-Based Media
FIGURE 13.8
The film Chariots of Fire won an Oscar for Best Picture in 1981. The movie included a one-minute scene, pictured here
with still images, of two athletes competing in the Great Court Run. The runner’s goal was to complete one lap around the
main college square before 12 strikes of the courtyard clock (roughly 43 seconds). The edited scene was comprised of
30 shots from multiple camera angles. Through continuity editing, the director was able to portray the scene in real-time
with continuous action, an amazing feat considering the scene was likely shot using only one camera. Notice how many
cut-ins are interspersed with shots of the runners.
Tech Talk
FIGURE 13.9
Starting a New Project in Adobe Premiere CS 5.5. (A) The editor is prompted to identify the scratch disk locations
for the main project (captured video and audio, project file, etc.). All of the media assets and project files associated
with a video project should be stored in a master project root folder. (B) Based on the information you provide in the
scratch disk window, the NLE will create physical folders and files on your drive associated with the project. (C) The
user is prompted to identify a video preset for the project sequence. Normally, you will choose a preset that conforms
to the format you shot your footage in. Is your project shot in SD or HD? Which codec and record settings were used?
(D) The program launches and editing can begin.
400 Section 4 Time-Based Media
FIGURE 13.10
Apple did away with the concept of “scratch disks” with its release of Final Cut Pro X. By default, the program now stores
media files in a master project folder called Final Cut Events and project files in a folder called Final Cut Projects. For
each new video project, the user creates a new event folder where media (source footage) and other assets associated
with the event will be stored. The project folder contains one or more sequences, the edited arrangement of media within
the NLE timeline.
an editing project. For example, digital video (DV) requires 1GB of storage
for every five minutes of recorded footage (or 250MB per minute). High-
definition video files consume an even larger amount of digital real estate,
which is why high-capacity hard drives are often used for storing the
contents of video and audio projects.
When a media file is imported into an active NLE session, a clip is created
and added to the media browser or project bin. A clip is a relatively small
data file that “points” to a larger underlying media file on the hard drive. It
functions as an alias or virtual representation of the real thing, and for this
reason, it is sometimes called a pointer file. A clip can be placed in the
timeline, trimmed, cut, and deleted without changing or destroying the actual
media file it is linked with. The same clip can be added to a timeline sequence
multiple times without adding to the physical size of the project file. Each
instance of the clip is a stand- alone digital asset that keeps track of the pointer
data for a particular event in the edited sequence. You can delete a media clip
from the bin without deleting the media file. While deleting it makes the clip
icon disappear from view in the project window or browser, the actual media
file is safe and sound in its original
FIGURE 13.11
Nondestructive Editing. Clips can be copied, deleted, moved, cut, and edited numerous
ways within the NLE without affecting the underlying media they refer to. If you delete a clip
from the project window, simply import it again into the program.
402 Section 4 Time-Based Media
location on the hard drive. A media file can be reconnected with a media clip at
any time by re-importing it into the project.
CAPTURE FILES
Capture files are time-based media files that are created when footage is ingested
into the computer through a connected camera or videotape recorder. The
video and audio tracks recorded on tape exist as linear streams of digital
information that are read and rendered in real-time as the tape is played back.
Capturing involves converting these time-based streams (or video and audio
tracks) into a digital file format that can be read and processed by the NLE
program. Once cap- tured, a media file can be easily moved and copied, allowing
it to be used in other projects or saved in a digital archive for retrieval at a later
time (see Figure 13.12).
Capturing-on-the-Fly
An NLE’s capture interface provides the editor with numerous options for
ingest- ing tape-recorded footage into the computer and saving it to the
project folder.
FIGURE 13.12
The Log and Capture interface in Final Cut Pro 7 allows users to ingest footage from tape via
a camcorder or VTR connected by a Firewire cable. The virtual player controls in the software
Time-Based Editing CHAPTER 13 403
interface can be used to operate a connected device.
404 Section 4 Time-Based Media
A virtual interface allows you to operate the playback and control functions of
a connected camera or VCR. The simplest method is called capturing on the fly.
With this process, the editor cues up the tape to the beginning, presses play,
and begins capturing. Capturing will continue until the editor escapes from the
procedure or the tape runs out of footage. The advantage of this method is that
it’s fast. A disadvantage is that it is fairly nonselective. The footage is captured
in large chucks—bad takes and all—and the clips are rather long and unwieldy,
which can slow down the editing process. For short projects with less than 30
minutes of footage, this method works well; but if you have lots of footage to
ingest, you should consider the more advanced method of logging and batch
capturing, which allows you to be more selective.
RENDER FILES
From time to time, the editing software creates new video files representing the
placement of a transition, effect, filter, title, or multilayer composite. This pro-
cess is called rendering. For example, when a one-second dissolve is attached
to two adjacent clips in the timeline, a one-second video clip will need to be
rendered showing the crossdissolve. A render file is a video clip created in the
background by the NLE whenever part of the underlying media file needs to be
copied, changed, or altered.
Time-Based Editing CHAPTER 13 405
FLASH BACK
NLE Pioneers
Avid Technology is one of the oldest developers of nonlinear editing software. Its
flagship program, Media Composer, was released in 1989 and continues to be
one of the most recognized NLE applications used in the motion picture and
television industries. Adobe’s NLE solution is called Premier, which Adobe
released in 1991.
Adobe’s chief competitor, Macromedia, was instrumental in developing a rival NLE
called KeyGrip, a product designed to work natively with Apple QuickTime. KeyGrip was
renamed Final Cut and eventually sold to Apple, who released it commercially as Final
Cut Pro in 1999.
While there are many professional NLE solutions to choose from, Avid Media
Composer, Adobe Premiere, and Apple Final Cut Pro have stood the test of time,
and remain three of the most recognized names in nonlinear video editing.
FIGURE 13.13
Avid Media Composer 6.
You can render clips as you go or wait until the end of a project to render every-
thing at once. The more complex the effect or transition, the longer it will take
for rendering to be performed. Depending on the speed of your computer, the
NLE can often play back unrendered clips in real-time. However, the resolution
may be low because the program has to process the transition or effect on the
fly. It’s a good idea to stop occasionally throughout a session to render all of the
unrendered clips in the sequence.
406 Section 4 Time-Based Media
The Timeline
Video and audio editing takes place in the timeline, a visual interface component
that’s divided into horizontal regions called tracks. When a video clip is
added to the timeline, the video portion of the clip is placed on a video track
while the audio portion is inserted directly below on an audio track. Audio
tracks are nor- mally grouped in linked stereo pairs. By default, the first track
(or left channel) is sent to the left speaker and the second track (or right
channel) is output to the right. For this reason, whenever a synchronized
video clip (such as an SOT) is placed on the timeline, its content is usually
spread across three adjacent tracks (one video track, and two audio tracks). If
a single microphone was used in the recording, there will be only one channel
of associated audio, and thus only one corresponding audio track. A mono
track can be easily converted into a stereo pair within the NLE, sending it to
both the left and right speakers.
When editing a synchronized video asset, you will normally want to apply
the same action, simultaneously, to both the video and audio portion of a
clip. For example, when a clip is split in two using the splice or razor blade
tool, the edit will typically affect all three tracks at once. Likewise, when you
reposition a video clip within the timeline, the linked video and audio segments
will travel together in tandem, allowing them to remain in sync. Problems can
occur when video and audio become unsynchronized. Even a slight shift of only
a few frames can affect lip-synching in an SOT, causing the audience to
squirm in discomfort.
Sometimes an editor will choose to unlink a clip in order to control the video
and audio separately. For example, you may want to perform a split edit, by
assigning different IN and OUT points to the video and audio clips. This
tech- nique is commonly used when editing sound bites. With a split edit, the
audi- ence might see a person talking for one to two seconds before they actually
hear him. Another use of this technique would be to show B-Roll during the first
five seconds of a sound bite before cutting to a synchronized headshot of the
person speaking. Split edits are performed routinely and are extremely
Time-Based Editing CHAPTER 13 407
versatile and important in the aesthetics of editing. Just remember to relink
the clips again
408 Section 4 Time-Based Media
FIGURE 13.14
Example 1: Creating a two-track composite image in Apple Final Cut 7. An SOT and lower third title graphic are added to
the editing timeline (bottom). After positioning the playhead over both clips in the timeline, we can look at the program
Time-Based Editing CHAPTER 13 409
monitor to see what the combined image will look like (top).
410 Section 4 Time-Based Media
composite image, showing a superimposition of the title over the top of the
person speaking. Because each component of the composite images resides
on its own track, it can be edited independently without affecting other clips in
the timeline.
Example 2: A Four-Track Composite. A color gradient background is positioned
on V1. Next, an SOT is added to V2 (see Figure 13.15). The SOT image is
scaled down to fit within the left side of the video frame, revealing the gradient
back- ground around it. A still graphic is added to V3 and positioned on the right
side of the frame opposite the person speaking. Finally, a lower third is inserted
on V4, showing the name and title of the person on screen. The composite image
shows all four elements co-residing within the image. Incidentally, the four
elements do not have to appear or disappear at the same points in the timeline.
The editor can choose a different IN and OUT point for each object. The
possibilities are endless.
Mixing is the audio version of compositing. With mixing, multiple audio clips
can be combined in an unlimited number of ways to form a complex aural
experience. For example, in a motion picture film, an action sequence often
contains multiple
FIGURE 13.15
Time-Based Editing CHAPTER 13 411
Example 2: Creating a four-track composite image in Apple Final Cut 7. This composite includes a background gradient, a
foreground video frame, a lower third title graphic, and a foreground program graphic.
412 Section 4 Time-Based Media
tracks of dialogue, background sounds, music, and sound effects (SFX). A sound
editor is one of many people responsible for ensuring that the audio
experience makes sense to the audience. Unfortunately, most of us don’t have a
sound editor standing by to perform the audio portion of video editing for us. We
will need to do this ourselves. And it needs to be done well, lest the audience is
overwhelmed by chaos and noise. Again, let’s consider a common example.
Example 3: A Four-Track Stereo Mix. We begin a sequence by placing an SOT in
the timeline. The audio portion of the SOT resides on A1 and A2 (see Figure
13.16). Music is then added to A3 and A4 and placed beneath the tracks of the
person speaking. In this example, the primary, or foreground audio, is produced
by the SOT. The editor wants this to stand out above the music and sets the
channel vol- ume (for A1/A2) to full. Music is a background element, and as
such should not compete for the audience’s attention. The editor adjusts the
channel volume (for A3/A4) to a lower level appropriate for background music.
As additional sources are added to the sequence, the editor will mix them
accordingly, until he or she achieves the intended balance and effect that the
program calls for.
FIGURE 13.16
Example 3: Creating a four-track stereo mix in Apple Final Cut 7. Combining audio sources in editing is a common task.
The editor needs to set the level of each stereo track pair independently to achieve an aesthetically pleasing mix. Since
the SOT is the main element, tracks A1 and A2 are set to normal levels (full). The music tracks are set much lower to
keep them from competing with the spoken word.
Time-Based Editing CHAPTER 13 413
Transport Controls
The virtual transport controls act in much the same way as the buttons on a VCR or
DVD player, allowing you to scan the contents of a time-based media clip. The
editor has access to familiar functions like play, pause, stop, rewind, and fast-
forward. The transport controls can be activated through keyboard input as well.
Experienced editors prefer using keyboard shortcuts (see Table 13.1) because
they’re generally much quicker to execute than the virtual controls. Shuttle and
jog are terms used to describe the speed at which you scrub (move) through
a time-based clip. The jog control allows you to move through slowly, a few
frames at a time, while the shuttle allows you to quickly scan through footage.
The Playhead
The playhead defines your position in a time-based clip. You can scrub slowly or
rapidly through a clip by dragging the playhead horizontally in either direction.
Audio scrubbing allows you to hear the audio portion of a clip as you move
the playhead back and forth, and it’s a useful technique for locating sound
artifacts that you want to eliminate or isolate in editing. The preview monitor
playhead allows you to scrub through the currently loaded clip. The program
monitor playhead allows you to scrub through the entire length of the
timeline. Because the program monitor is linked directly to the timeline,
moving either playhead causes the other one to move in tandem.
Timecode Fields
Every frame of video is addressable via timecode. Whenever you mark an IN
point or OUT point, the location is stored as a numerical series of eight digits
denoting hours, minutes, seconds, and frames. A timecode display of
01:18:54:27 indi- cates a frame location of 1 hour, 18 minutes, 54 seconds,
and 27 frames. This format is used to denote the duration of marked clips,
transitions, and other program assets. For example, a one-second dissolve
would appear as 00:00:01:00 in the duration field.
414 Section 4 Time-Based Media
GREAT IDEAS
NLE Keyboard Shortcuts
FIGURE 13.17
The source or preview monitor is used for screening and trimming clips before adding them to an edited sequence. The
original clip shown here is 3:44, but the editor only wants to use a portion of it in the final edited sequence. He marks
the portion of the clip he wants to use with an IN point and an OUT point. When the edit is performed, only the video and
audio between these two points is added to the timeline.
416 Section 4 Time-Based Media
In a video editing timeline, the audio portion of the narrative is the structural
frame that everything else attaches to. It needs to be created first. Doing so
allows you to edit video “in time” with the dialog, narration, music, and other
audio elements in the sequence. The beat, tempo, and rhythm of the audio
cues the editor to natural points in the timeline where transitions make the
most sense. It is very difficult to edit video images without audio in the time-
line. Once the audio portion of the narrative is established, editing the visuals
becomes a snap.
FIGURE 13.18
The three-point editing technique is used here to insert a 30-second portion of a 3-minute clip
at a specific point in the timeline.
Time-Based Editing CHAPTER 13 417
I scrub the playhead to the beginning of VO #1 (the best take) and mark an
IN point by pressing the I key on my keyboard. Next, I move the playhead to
the end of VO #1 and mark an OUT point by pressing the O key on my
keyboard. The segment is now marked and ready to be placed in the timeline.
The marked clip is 30 seconds long (00:00:30:00).
Transitions
A transition is an effect applied to the timeline at the beginning or end of a video
or audio clip. Transitions are used to provide a sense of flow and rhythm to a
project and to guide the viewer’s senses through changes in the narrative
struc- ture of a story as it unfolds (see Figure 13.19).
CUT
The most common transition is called a cut. A cut is an instantaneous
transition from the end of one shot to the beginning of another shot. Because
of this, a cut has no duration per se. In fact, you can think of a cut as the
invisible transition. Cuts are popular with editors because of their simplicity.
When properly timed, the audience will hardly know that a cut occurred.
Timing is critical, though.
FIGURE 13.19
(1) The first sequence (top) illustrates the visual effect of a 1-second dissolve transition at the midpoint between the
end of one shot and the beginning of another. (2) The second sequence (middle) shows a timeline view of the attached
dissolve. Transitions are usually attached to clips by dragging and dropping them into place on the timeline. (3) The third
sequence (bottom) shows a few of the transitions available in Final Cut Pro X. While cuts and dissolves most often suffice,
your NLE software will provide you with many basic transitions to choose from.
Time-Based Editing CHAPTER 13 419
Cuts should be positioned at natural points in the narrative; for example, at the
beginning or end of a sentence or concurrently with a sound effect or music cue.
Cutting in the middle of a word or sentence, or on the wrong bar or beat in the
underscore, can disrupt the continuity of a program and be self-defeating. Most
often, the editor’s goal is to make the presence of a cut invisible to the audience.
Remember, the best edits are often the ones you don’t notice. Incidentally, the
first beat of a bar, called the downbeat, is the strongest point in a melody line
and a great location for the placement of a cut or other transition.
FADE
The absence of picture data on a television screen is referred to as black. In
televi- sion, the audience is cued to key program and scene changes by fading
into or out of black, in much the same way that a lighting blackout is used
during a the- atrical production to signal the beginning or end of a scene or act.
Programs usu- ally begin with a fade-in from black and end with a fade-out to
black. The fade-in is a gradual transition from black to a fully opaque
television image, while a fade-out is a gradual transition in the opposite
direction. Video and audio fades often occur simultaneously. For example, in
an end-of-program fade-to-black, the video image should reach 100% black
at precisely the same point in the timeline that the audio reaches 0 dB. In a
case like this, it is standard practice to use the same duration time for both
the video and audio fade.
DISSOLVE
A dissolve is a gradual transition from one shot to another that is created by
overlapping the fade-out of one clip with a fade-in of the next clip in
sequence. Dissolves are far less abrupt than cuts and are often used to signal
a change in time, a change of location, or a change in tempo. Dissolves can
be used to slow down the pace of a program when a gentler timing structure
or rhythm is called for. However, short dissolves can be used in place of a cut
to soften a transition without necessarily impacting an otherwise upbeat
tempo.
WIPE
Most other transitions fall into the category of wipes or 3D effects. A wipe uses
a linear movement or sweeping pattern to replace the current image with a
new one. For example a vertical wipe moves from the top of the frame down,
or from the bottom of the frame up, as it replaces the current image with new
pixels. A horizontal wipe does the same thing by sweeping left to right or right
to left. A circle wipe moves in a radial direction from the center of the image
outward and vice versa. Wipes come in a virtually unlimited assortment of
shapes and patterns. There are checkerboard wipes, clock wipes, and slide wipes,
which push the entire frame of video off-screen. A spin wipe is used to rotate
the video frame as it moves out of view, while an explosion wipe shatters the
picture into pieces before sending them off in every direction. If your selec-
tion of wipes is limited, chances are that a third-party software developer has
produced wipes for your NLE that can be obtained for free or purchased and
downloaded commercially.
420 Section 4 Time-Based Media
Like wipes, 3D transitions come in all shapes and sizes. Using powerful
algorithms, video can be squished, stretched, warped, morphed, and distorted
in any num- ber of ways to achieve creative and unusual visual effects. Page
peel transitions curl the edge of the video frame inward to simulate turning
pages in a book. Cube spin transitions turn the video frame into a rotating
object with multiple sides displaying different images or picture data. The list
goes on and on.
Wipes and 3D transitions are visually very cool! They are fun to play with
and can add legitimate value to a visual experience for the viewer. However,
you must use them with caution. Anything you add into the video timeline
should be motivated by the story and the impact you wish to make on the
viewer. Sim- ply throwing things into an edited sequence because you can, or
because it looks cool, is not necessarily going to make it look any more
professional. In fact, it’s usually just the opposite. Until you gain experience,
stick with cuts and dissolves for the most part. If you choose to use wipes,
then be consistent. Do not use a pinwheel wipe in the first transition, a page
peel in the second, and exploding video in the third. This sort of sporadic and
random use of transitions will likely overwhelm the viewer’s senses,
potentially causing them to lose interest and disengage from the message.
TRANSITION PROPERTIES
Once a transition has been added to the timeline, it becomes a modular object
with properties that can easily be changed. These values are saved in a data
file that’s stored in the render folder. While most transitions will play back in
real- time, at some point before exporting your project, you will need to stop
and fully render all of the transitions you’ve created. Once rendered, they
will run more smoothly and appear sharper when playing through the video in
real-time.
The most important property you will need to pay attention to is the
transition duration. Duration is the amount of time that a transition takes to
perform from beginning to end, as indicated in seconds and frames. The
default transition in most programs is one second (or 30 frames). When
adding a dissolve between two clips, I often use a 20-frame duration, which
is roughly two-thirds of a sec- ond. This duration serves me well most of the
time. For something heavily dra- matic, I might use a longer increment of 1
to 3 seconds. Only rarely, will I use anything longer. You need to remember
that transitions take time to execute and that this eats into your project’s total
runtime; robbing you of precious seconds and minutes, especially if the
program is short. For example, if you produce a 60-second commercial with
10 default transitions (01:00 long), your program loses 1/6th of its content
potential to transitions. The simple act of changing the default duration to 15
frames will save you 5 seconds of actual runtime. Five seconds in a 60-second
spot is a lot of time.
In addition to the duration property, you can adjust a number of other vari-
ables. For example, you can change the direction of an edge wipe, causing it
to start from the left, right, top, or bottom of the frame. The properties you
have access to will depend on the transition you’re working with. Each
Time-Based Editing CHAPTER 13 421
transition is
422 Section 4 Time-Based Media
unique, and therefore has different parameters and values that you can adjust
or manipulate to change the way that it performs.
When applying a transition, make sure that the adjoining clips you’re
attaching it to touch. If there’s a gap between the two clips that you can’t see
—because you’re zoomed out too far, for example—then the transition may
not attach properly. Zooming in to the timeline for a close-up view of the
attachment point is really helpful. Once attached, transitions become visual
objects you can mod- ify. For example, you can change the duration of a
transition by entering a new value in its properties window; or you can do it
interactively in the timeline, by dragging the left or right edge of the
transition in or out until the desired dura- tion is obtained.
Not surprisingly, you’ll find that the interfaces for audio and video editing change
CHAPTER
between SUMMARY
programs. What doesn’t change, however, are good editing practices. For
example, the need to use cutaways correctly and avoid unintentional jump cuts doesn’t
change when you switch from one NLE to another, nor does the need to carefully
monitor your audio. In addition, you’ll find that many of the core concepts are the same
between NLEs. Most professional programs use a timeline for arranging clips linearly
from begin- ning to end, allow you to organize your clips in bins, and mark in and out
points in your video. The basic idea of a three-point edit is the same, regardless of
your NLE. Key- strokes may even translate between NLEs, particularly the use of the
space bar to start and stop the playhead. While you need to be able to use your NLE
effectively, don’t get too tied up in the software—if you don’t understand basic editing
concepts and theory, you won’t be able to communicate effectively. Try to get
experience on as many NLEs as possible, and avoid becoming too attached to a
particular NLE. When you are looking for a job, you want to be as flexible as
possible.
Time-Based Editing CHAPTER 13 423
Backslash character, 59
Browsers, 188, 401 CNN iReport, 17f
Backspace, 52
Buttons, 208 Code checker, 210–212
Balance, 121–122
Bytes, from bits to, 36–37 Codecs, 325–326
Balanced microphone cable, 343
Bytes-per-second, 61 Collaborative filtering, 173
Balsamiq, 85f
Color, 113–114, 120
Bandwidth, 68, 375
Baseline, 250
C background and text,
Cable Management, 346 206 depth or bit depth,
shift, 267
Camcorder, 359, 362 225 model
Basic tag, anatomy of, 186f CMYK, 224–225
invention of, 361
Bass frequencies, 320
large-format/small-format, 365f RGB, 224
Batch capturing, 399
Camera-recorder, see Camcorder sampling, 372
Batch computing, 40
Canon XL-1, 367f space, 224
Batch processing, 30
Canted angle, see Dutch tilt Color temperature, 110
Batch Rename tool, 30f
Canvas, 101, 103f .com, 189
Benchmarking, 88
Capacity of hard drive, 60–61 Command key, 53
Betacam, 362, 362f, 365
Capline, 250 Command line interfaces (CLIs), 49, 155
formats, 366
Caps-Lock key, 53 Communication, 9
Betacam SX, 365
Capture files, 398 with binary numbers, 38
Betacam-SP and MII, 362–364
Capturing, 399 Compact disc (CD), 10, 63–64,
Betamax, 372–373
Capturing-on-the-fly, 398–399 324 Compact disc read-only
Beveling, 274
Card reader, 65 memory
B-frame, 369
Cardioid microphones, 331f, 332 (CD-ROM), 63
Bidirectional microphones, 331, 331f
Carousels, 164–166, 166f Compact Flash (CF), 307
Binary
Cascading style sheets (CSS), 179, Compatible Time Sharing System
communicating with, 38–
194, 194f, 204–205 (CTSS), 40
39 decimal to, 36t
box model, 146, 146f Complex tag, anatomy of,
Bit depth, 225, 323–325
Catch Me If You Can movie, 186f Composite video system,
Bitmap imaging, 220
114 Cell padding, 204 363 Compositing, 402
Bits, 22, 36
Cell spacing, 204 Composition, 101, 119
to bytes, 36–37
Cells, table, 142, 144–145, 145t balanced, 121
Bits per second (bps), 61
Celluloid, 23 graphic, 106
Bitstream, 325
Centered text, 271 Compression, 230
Blank spaces, HTML,
Center-weighted metering mode, artifacts, 231
187f Block quotes, 197
294 Central processing unit (CPU), codec, 325
Block-level elements, 185
43, phases, audio, 319
Blogging engines, 4
45, 45f Computer storage, 56
Blue laser disc, 64f, 65
Channel, 401 Computer system
Blu-ray format, 374
Character keys, 52, 52f components of, 43f
Body, HTML, 184
Chariots of Fire movie, 393f hardware, 44–49
Body copy, chunking, 133
Checkboxes, 170, 171f processor speed, 45
Bold typeface, 250
Chip manufacturers, 45 system memory, 46–49
Boom pole, 337
Chroma subsampling, Concept development, 77–88
Booting up, 43
372 Chunking body copy, Condensed typefaces, 262–263
Borders, 204
133 Circle wipe, 411 Condenser microphones, 328–329,
table, 145
Client, 76–87 328f, 330f
Boundary microphones, 339
Client meeting, initial, 76– Consumer
Box layout, 194f
77 Client-side script, 209 cameras, 278–279
coding the, 195f
Clipping, audio, 349 formats, 372–373
Box model, CSS, 146, 146f
Clips, media, 396–400 as producer, 16–17
Brainstorming, verbal, 79
CLIs, see Command line interfaces Consumer-generated media (CGM), 16
Breadcrumbs, 162–163, 162f
Clock speed, 45 Container formats, 325–326, 375
Brightness, 113, 363
Closing tags, 183 Content, 96–99
Broadcasting, 19
Cloud storage, 67, head, HTML, 184
B-Roll, 388, 401, 409
67f Clutter, avoiding, headings, 197
over VO segments, 409 391 hyperlinks, creating and
CMS, see Content management managing, 198–201
system CMYK color model, 224–225
Index 417
8-Core processor, 45
File size, 289–290 Fox News, 18f
EDTV, 242
File system browser, mastering, 56, Fragmented hard drive, 62
Electronic media, see
56f File Transfer Protocol (FTP), 190 Frame, 102, 112, 116, 121, 238
Time-based media
File-base recording formats, 374–382 Frame rate, 240
Electronic newsgathering
open standards for, 375–377 Frequency, 317, 319–321
(ENG), 361, 365
proprietary standards for, 377–382 F-stop, 286–287
Electronic Numerical Integrator and
Fill flash, 294 Full-screen graphics, 389–390
Calculator (ENIAC), 40
Film, photographic Fully automatic modes, 302
Element, 183, 186
size, 288–289 Function keys, 53–54, 54f
block-level, 185
speed, 288 Functional customization, 173
Elements of design, 101–114
Film recording, motion
Email link, creating, 200
Emboss, 274 picture, 353–356
Final cut, 76
G
Emphasis, 118–121 Garalde typefaces, 253
boldface for, 262 Final Draft scriptwriting program, 83
Geometric shapes, 108
principles of, 114 Firewire, 68, 69f
Geometric typefaces, 254
400 port, 69
End key, 53 Gestalt psychology, 115
ENG, see Electronic 800 port, 69
Gestalt terminology, 124
newsgathering Englebart, Douglas, Firmware, 42
Gigabyte (GB), 56
49, 50f Enhanced definition Fisheye lenses, 284
Golden ratio, 135, 135f
television Fixed layouts, 150–151
Graph paper, 137, 138f
(EDTV), 242 Fixed storage, 59–62
Graphic vector, 124, 124f
Enter key, 52 Flash control, 294–295
Graphical user interface (GUI),
Escape key, 53 F-keys, 53
49–51, 155
European Organization for Flash memory, 65–66 Graphics, 48, 217, 218f
Nuclear Research (CERN), F-layout, 129–131, 130f bitmap, 220
182 Flicker fusion threshold, 240 card, see Video card
Evaluative metering mode, 294 Floating graphic, 141, 142f full-screen, 389–390
Exabyte (EB), 56 Floppy disk, 66f vector, 233–235
Expanded typefaces, 262–263 Fluid layouts, 151 Graphics processing unit (GPU), 48
Exposure, measuring light Fluorescent, light source, 296 Great depth of field, 300
intensity, 294 Focal length, 302 Grid, 105, 271
Exposure meter, 294 Focal point, 123 rule-of-thirds, 123
Exposure modes primary, 118, 124 scale, 137
automatic modes, 302–303 visual, 114 system, 136–142, 145t
manual modes, 306 Focus graph paper, 137,
semi-automatic modes, 304– autofocus, 297–299 138f modular, 142
306 manual focus, multicolumn, 141
Expression Web, 183 299 Folders, 58f parts of grid, 139,
External bus, 68 Font, 207 139f
External hard drive, 59–60, 60f color, 266–267 single-column, 140, 143f
External microphones, 340– sizes, 264–266 structures, 139–142, 140f
341 External sites, linking, 198f, styles, 261–267 typographic, 136, 139f, 142
200 External style sheet, 204 transformations, 272–274 Group of pictures (GOP), 369
Eye candy, 98 Font family, 246, 247f GUI, see Graphical user interface
Font management software, Gutenberg diagram, 129, 129f,
F 268 Footage, 383, 386 140 Gutenberg press, 247, 247f
Facebook, 6, 150 Footers, 163–164, 163f
Foreground, 124
Fade-in transition, 411
Fade-out transition, 411 sound, 388 H
Formal balance, 122 Halftone image, 105, 105f
Fat footer, 163, 163f
Forms, 109–111, 169–172, Handheld microphones, 332–334, 334f
Faux font styles, 261
171f, 172f Hard drive, 43
Favicon, 186, 233
Forward slash, 59 connecting with devices, 68–69
Field of view, 102, 102f
Four-track composite, 403 maintenance and optimization,
Field, raster scanning, 239
Four-track stereo mix, 404 62
Figure-ground, 124–125, 125f
Hard lighting, 110
File management, 58–59
Hard-disk drive (HDD), 59
File naming conventions, 191
Index 419
Hardware, 42 iLink, 68
Image folders, 309
computer, 44–45
Image(s), 218
HD video standard,
bitmap, 220
151
background, 206
HDCAM, 364f, 365–366
downloading, 308–309
HDTV, 242
floating, 141
HDV, 368–369
frame, 406–407
Head content, HTML, 184, 186
linking and adding, 201–202, 201f map, 200
Headings, 133, 134f, 143f
moving, 218 photographic,
content, 197
219f raster
Headline, 133
anti-aliasing, 228–230 definable element
Headphones, 350–351
of, 226–227 formats, 232–233
Helical scanning, 357
pixels, 221, 222f
Hertz (Hz), 238
resampling, 228
Hi-8 analog tape formats, 362f
Hierarchy, establishing, 192, 193f resolution, 223, 223f
High-contrast background, 391 scaling, 227–228, 228f
High-definition tape formats, renaming, 309
366 High-definition video files, rendering, 237–239
397 High-key image, 119 sensor, 288
High-speed USB, 68 separate, 200
Home key, 54 stabilization, 306–307
Home links, 162 vector, 220 Imaging
Horizontal lines, 106 chain
Horizontal wipe transition, iris, 285–287
411 lens, 281–285
Howdy Doody children’s program, shutter, 287–290
354f HTML, see Hypertext Markup image sensor, 288
Language HTML 5, 182, 184f, 185, IN point, 385
209 Index vector, 124, 124f Index.html,
Hue, 113, 363 191
Human frequency spectrum, Informal balance, see Asymmetrical balance
320 Human interface, 49 Infrared transmitter, 342
Human interface devices (HIDs), 49 Inline element, 185 Inline graphic,
Humanist typefaces, 254 141, 142f Inline styles, 204
Hypercardioid microphones, 331f, 332 Input prompt, 171, 172f
Hyperlinks, creating and Institute of Electrical and Electronics Engineers
managing, 198–201 (IEEE) 1394, 68, 69f
Hypermedia, 10, 10f Intel 8088, 46
Hypertext, 10 Intel 8080 microprocessor, 40
Hypertext Markup Language Intellectual property, 75
(HTML) blank spaces and line Interactivity, 154
breaks, Interframe compression, 369
186, 187f Interlaced scanning, 238–239
block-level elements, 185 Intermediate formats, 377
Dreamweaver and Expression Internal bus, 68
Web, 183 Internal hard drives, 59–60, 60f Internal style
head content, 186 sheet, 204 International Electrotechnical
HTML 5, 182, 184f Commission (IEC), 56, 374
lists, 197
nesting tags, 184,
185f sections of, 184
Hypertext Transfer Protocol
(HTTP), 189
I
IBM PC, 41
Idea generation, 77–88
420 Index
International Organization for
Standardization (ISO), 374
Internet, 14, 20t
Explorer, 56
history of, 182
Intraframe compression, 367, 369
Intranets, 190–191
IP address, 190
Italic typefaces, 250
J
JavaScript, 208
Jog control, 405
Joint Photographic Experts Group
(JPEG), 230
Jump Cut, 393
Jump drives, 66
Justification, 116, 270–271
K
Kanizsa’s Triangle, 126f
Kay, Alan, 41f Keyboard
basics, 51–55
shortcuts, 54–55
KeyGrip, 400
Keying, 388
Keystroke combinations, 54
Kilobyte (KB), 56
Kinescope, 353, 355, 355f
Kodak Corporation, 277
L
LAN, see Local area network
Lands, optical disc, 63
Landscape mode, 302–303
Language of New Media, The
(Manovich), 23
Large audience, 18–20
Large font sizes, 391
Large organizations, product of, 15–17
Large-format camcorders, 365f
Lavaliere microphones, 335–336 Law
of
continuation, 124
proximity, 115
similarity, 116 Layout,
see Page layouts Leading
lines, 124
Learnability, 175
Legibility, 248–249, 248f
Lens, focal length of, 302
Letterform, 246
postures, 250
proportions and, 250–251
Index 421
News aggregators, 5
dynamic, 147 Plosives, 333
Niche audience, 20
layouts, 127 Plug-in, 188
Nielsen, 18
fixed, 150–151 Pointer file, 397
Night mode, 303
F-layout, 129–131, 130f Polar pattern, 329–332
1996 Telecommunications Act, 179
fluid, 151 Portrait mode, 302
Nintendo 3DS, 155, 156f
page design and, 193– Positive space, 104
Nonbreaking Space, 187
194 Z-layout, 131–132, Postproduction, 75–76
Nondestructive editing, 397f
131f Posture, 250
Nonlinear editings (NLEs),
organizing content on, Power supply, 43
383–385, 394
127 peel transition, 412 PowerPoint, 27, 28
interface, components of
splash, 147 Preamp setting, 348
preview and program
static, 147 Preproduction, 74
monitors, 405–407
structure and title, 195–196 Pressure zone microphones
project window, 401
templates, see Templates (PZM), 339
timeline, 401–404
Palo Alto Research Center (PARC), Preview monitors, 405–407
pioneers, 400
49 Panasonic D5, 366 playhead, 405
programs, 363f
Panasonic D3 tape format, Previsualization tools, 78–87
Nonvolatile memory, 43, 46
365 Panasonic D5HD, 366 Primary optical area, 129
Normal lenses, 284
Panasonic M tape format, 362 Principles of organization,
Normal writing conventions, 264
Pandora Radio, 20, 21f 115 Printing technology, 248
Numerical representation principles
Paradigm shift, 15 Processor speed, 45
of new media, 24–26
new media, 16–17, 20 Producer, 76–88
O Paragraphs, 197
Parent folder, see Subfolders
consumer as, 16–17
Production, 74
Oblique typefaces, 250 Passive autofocusing system, 297 Professional cameras, 279
Obtrusive observation Pattern, 112–113 Professional Disc, 377
method, PCM, see Pulse-code modulation Professional microphones, 343f
176, 177 Pentium III processor, 46 Professional scriptwriting
Old media, characteristics of, 14, 16f Perceptual forces, 121–125 program, 80
large audience, 18–20 principles of, 114 Program mode, 306
large organizations, 15–17 Persistence of vision, 240 Program monitor, 405–407
simultaneous delivery, 20–22 Personal computer (PC), 40–
Omnidirectional microphones, Progressive scanning, 238
42 Personalization, 172–173 Project file, 396
330, 331f Perspective, 107
Open standards, 374 Project folder, 394
Petabyte (PB), 56 Project window, 401
for file-base recording formats, P-frame, 369
375–377 Property, 200
Phantom power, 329 Proportion, 121
Opening tags, 183 Photo gallery, 29f
Open-reel tape recorder, 356, 356f, Proprietary card formats, 378, 379f
Photo-editing tools, 43 Proprietary standards, file-base
361 Open-reel videotape formats, Photographic film, 119
360 Operating system (OS), 41 recording formats, 377–382
Photography ProRes 4:2:2, 377
Optical disc formats, consumer cameras, 278–
373 Optical drive, 43 Prosumer cameras, 279
279 prosumer and Prototype, 86–87
Optical image sensor, 218 professional
Optical storage devices, 63–65 Proximity, 115–116, 115f,
cameras, 279 118f effect, 334
Optical zoom, 285 video cameras, 280
Option key, 53 Pseudo-classes, 208
Photoshop Express, 43 Psychological closure, 125
Organic shapes, 108 Pistol grip, 337, 338f
OUT points, 384 Pulse-code modulation (PCM),
Pitch 322, 392
Over-under wrap, 347 sounds, 321 PZM, see Pressure zone microphones
track pitch, 359
P
P2, 378, 379f
Pits, optical disc, 63
Pixels, 104, 221,
Q
Padding, 146 Quad recorder, see Quadruplex
222f
Page recorder
count, 226
design and layout, 193–194 Quad-core processors, 45
dimensions, 226
Placeholder text, 146, 149f
Playhead, 405
Index 423
T
165t Thunderbolt, 69, 69f
TIFF format, 230
U
Ultracardioid
Tab key, 52 Time-based media, 289f microphones, 331f, 332
Tables, 142–145, 144t, 145t Timecode, 405 U-matic ¾-inch, 361–362
borders, 145 Timeline, 401–404 U-matic SP, 362
cells, 142 adding clips to Underlining, 264
splitting and editing scenario, 408–410
Unidirectional microphones, 332
merging, 144–145, transitions, 410–413 Uniform resource locator (URL),
145t Time-shifting, 372 188–189, 189f, 198
creating, 202–203, 203f Tip, ring, and sleeve Unity, 114, 114–117
defining, 144 (TRS) connectors, Universal lighting, 111
Tabs, 160–161, 161f 344
Universal Serial Bus (USB), 65, 68,
Tag clouds, 168, 170f Tip and sleeve (TS)
69, 69f
Talbot, William Henry Fox, connectors, 344
microphones, 342, 342f
105 Talking head, 387 Titles, 388–389 Unobtrusive observation
Tape formats, 364f safe, 392 method, 176, 177
analog, 360–364 Tonal range, 119
digital, 364–365 Unscripted action, 386–389
Tonality, low-key, 119 Unscripted dialogue, 386–389
Tape recording, 359 Touch interfaces, 155 uReport iPhone app, 18f
Tape-based recording methods, 375 Tracks, 61, 375, 401 U.S. Broadcast TV Ratings, 19t
Tape-to-tape editing, 384 pitch and speed, 359
of videotape, 358,
358f
Index 425
Usability
compositing, 402–404 Web 3.0, 12
ensuring site, 195
compression, types of, 368– Web page
testing, 90, 213
372 editing, 383 background, 120
USB, see Universal Serial Bus timeline, 408 design template,
User interfaces, 153–158
recording, 357 118f
accessibility of, 178–180
sharing tools, 5 Web parent companies, 20t
CLIs, 155
track, 401–402 Web user interfaces (WUIs), 155
designing, 158 Webdings font, 256f
transcoding, 354
efficiency, 175 Website, linking, 200–201
Video card, 43, 48f
GUIs, 155 Video cassette recorder (VCR), Wertheimer, Max, 115
motion tracking, 157 357f U-matic, 361 White balance, 295–297
navigation, 159–164 Video Home System (VHS), White space, 104, 104f, 116, 121, 392
managing hierarchy, 162–164 372–373 Wide shot, 393
menus, 159–160, 159, 160 Videocassette tape formats, 360f Wide-angle lenses, 283
tabs, 160–161, 161f Video-friendly colors, 391 Widescreen, 237
organizing content of, 164– Videotape, 5, 358 Wiki, 5
168 pagination, 167–168, 167f formats, professional, 359–368 Windows Explorer, 56
tailoring, 172–173 recording, 356 Windows systems, 59
touch, 155 tracks of, 358, 358f Wipe, 411
types of, 155 Videotape recorder (VTR), 360 Wireframe, 83–84, 85f–87f, 406
usability, 174–178 invention of, 357 Wireless microphones,
and accessibility, 180 type C, 359 341–342, 341f
feedback, 177, 178 Virtual monitor components, 405 WordPress.com, 150f
testing, 176 Virtual Private Network (VPN), Workspace, 101
WUIs, 155 191 Visual communication, 99– digital, 102
User-centered design, 155–158 100 World Wide Web Consortium
User-generated content (UGC), 16 content and form, 96–99 (W3C), 182
Visual elements, 112, 112f, World Wide Web (WWW), 11–
V 121 12, 11f, 181
Value (contrast), 113, 118–120 Visual hierarchy, 132–133, 132f, 141 history of, 182
Variability, principles of new Visualization, 78 WOW-factor, 98–99
media, 31–32 Vlogging, 4 Wright, Charles, 15
Variable focal-length lens, 285 VO, see Voiceover Write once, read many (WORM), 64
VCR, see Video cassette recorder Voiceover (VO), 388 WUIs, see Web user interfaces
Vector graphic, 220, 234 clips, 409
Vectors, 124, 124f
imaging, 220
Volatile memory, 43, 46 X
Volume-unit (VU) meter, 348 XDCAM, 377, 378f
Verbal brainstorming, 79 VTR, see Videotape recorder XLR connectors, 343f, 344
Vertical lines, 107
Vertigo, 108 W Y
VHS analog tape format, 362f Warm colors, 120
Video YouTube, 154f, 173
Warping, 272
adapter, see Video
card blogging, 4
Waveform, 406
Waveform audio file format
Z
cameras, 280 (WAV/WAVE), 325–326 Z-layout, 131–132, 131f
Web 1.0, 11
Web 2.0, 11