| by Stephanie Argy for The
                    Motion Picture Editors Guild Magazine
 
                    
                      | "How
                        High" — a medium-budget comedy, is one
                        of the first studio features to be shot on 24P video,
                        using the Panavision/Sony 24P camera. The picture features
                        rap stars Method Man and Redman playing themselves and
                        was directed by Jesse Dylan (Bob Dylan’s son). In this
                        interview, editor Larry Bock and assistant editor Erik
                        C. Andersen cover their experience editing the show: why
                        it was produced this way, what it was like to work with
                        the new medium and how it affected the post-production
                        process. 
 |  | 
                          
    |  |  
    | For a preview, the production installed a high-definition projector in the theater balcony. Standing: Robert Jacobs, tape operator from Video Applications, which provided the projector; Andersen; Bock; John Haley, audio engineer, Video Applications. Seated: John Woo, assistant editor; Keith R. Brown, editorial production assistant; Daniel O’Keeffe, projection engineer at Universal. 
 |  |  How did you prepare to work in this new medium?
                   Bock: I first called my friend Michael Alberts,
                    who recently finished cutting a low-budget film called Nicolas,
                    the first feature shot with the Panavision/Sony 24P high-def
                    camera. He walked me through the process and alerted me to
                    certain problems we might expect. Then I called Erik, whose
                    technological knowledge spans a wide spectrum of different
                    areas in post. Andersen: I read everything available on
                    24P, including the articles in the Guild Magazine. I also
                    spoke to a couple of assistant editors who had worked with
                    the format.  Bock: We also attended the camera and sound
                    tests for How High at Panavision. I was worried because it
                    seemed like there were so many opportunities for something
                    to go wrong with sync or timecode. But things went very smoothly,
                    thanks to the efforts of the film crew and my team. What were some of the advantages and disadvantages
                    of working this way?  
                    
                      | Creatively, working in hi-def was
                          no different than working on a film show. In general,
                          I was quite surprised at how much the show looked like
                          it had been shot on film.— Bock
 |  Andersen: Director Jesse Dylan and director
                    of photography Francis Kenny had to contend with the fact
                    that the Panavision camera didn’t have a full set of
                    lenses. [At the time How High was shot, there were three zooms
                    and one prime available; as of this writing, there are four
                    zooms and five primes.] The camera cannot be under- or over-cranked,
                    and it has contrast problems in bright sunlight and signal-to-noise
                    issues in low light. The viewfinder is black and white, with
                    very low image quality, but there is a color HD monitor on
                    hand to check focus, lighting and actors’ performances.
                    I was very impressed with the image quality on the monitor.
                    What you see is what you get. One big advantage to using the
                    24P camera is that HDCAM tapes last 50 minutes (at 24 frames
                    per second), much longer than film magazines, so the DP was
                    able to keep the camera rolling between takes and change magazines
                    less often. Production thought this saved roughly an hour
                    a day and allowed for more setups, but it meant that we were
                    getting a lot of unslated material that had to be sorted out
                    in the cutting room. Were any other high-definition cameras considered?
                   Andersen: They looked at the Panasonic AJ-HDC27A
                    720P DVCPRO HD camcorder, which has variable speeds of 3,
                    6, 12, 24, and 72 frames per second, but they were concerned
                    about mixing 720P images with the 1080P material produced
                    by the Panavision/Sony camera. It would have also have meant
                    a second camera package rental and a new set of lenses for
                    the DP. How did shooting on 24P affect your interaction
                    with the crew?  Bock: For me it was the same as always —
                    no different than shooting on film.  Andersen: I talked to the sound and camera department daily.
                    One issue was pre-roll. The 24P camera records time-of-day
                    code, which means there is a timecode break between each take.
                    There were many times I was unable to digitize a slate because
                    I only had a second of timecode before the director yelled,
                    “Action.” I needed ten seconds of code and often
                    didn’t get it. What was the workflow from the set to the editing
                    room?  Andersen: Hollywood Digital downconverted
                    the HDCAM masters to Beta SP. They transferred the full 16:9
                    image letterboxed and added two timecode windows in the matte
                    area: 29.97 timecode from the beta, and 24P (23.976) timecode
                    from the master. The Beta was digitized into the Avid and
                    we used audio from the Beta. Did you record audio on DAT as well as video?
                   Andersen: Yes. Cameron Hamza, our sound
                    mixer, hard-wired audio to the camera via 300 feet of specially-built
                    cable. The sound went from the microphone to the mixing panel;
                    then four tracks ran out of the mixer: two to the camera and
                    two to a Fostex PD-2 DAT. On 35mm, the camera, sound and video playback departments
                    work independently, but with the 24P camera, they are hardwired
                    together with sound, video and timecode cables. We added two
                    additional people to the film crew, a hi-def engineer and
                    an extra sound assistant to wrangle all the cable.  
                    
                      | We received an average
                        of three-and-a-half hours of material a day. Because there’s
                        no telecine, one big change for the assistants was that
                        we had to log the footage ourselves. 
 — Anderson
 |  One issue was how to record sound so that camera and DAT
                    timecode would match. Production used an Evertz Afterburner,
                    which produces a downconverted image and also converts the
                    camera’s 23.976 timecode to 29.97, which was recorded
                    on the DAT. Cameron Hamza used the Fostex DAT because it was
                    one of the only field recorders that could lock to an external
                    video sync pulse, which came from the Evertz. We thought about
                    recording audio timecode at 23.976 but decided not to risk
                    incompatibilities with post-production gear. The Afterburner
                    has a processing time of about five frames. So our DAT timecode
                    had a consistent offset relative to video, which was easy
                    for the sound department to fix in post. As a test, I digitized from the DAT, then synced in the Avid
                    via the slate. After accounting for the five-frame offset,
                    I found an additional one-frame audio advance. We discovered
                    that video processing in the camera delays the image by a
                    frame as it is being recorded. The camera is actually out
                    of sync with itself! To address this, Hollywood Digital delayed
                    the sound by a frame when they downconverted our dailies to
                    Beta SP for cutting. How did cutting room procedures differ from
                    those on a film show? Bock: Since we didn’t have to wait
                    for negative to be processed, Erik was able to digitize the
                    footage and I was able to cut the scenes within 24 hours from
                    the time they were shot. On film, we would have been one or
                    two days behind camera. John Woo was the second assistant
                    editor, and Jessica Caggiano was our apprentice. After principal
                    photography we let go of the apprentice. It was the smallest
                    crew I ever had on a feature.  Andersen: Because there’s no telecine,
                    one big change for the assistants was that we had to log the
                    footage ourselves. We used Avid MediaLog hooked up to a Beta
                    SP deck. John logged everything; in case the director asked
                    for a non-circled take, I could load it immediately. John
                    would log the first tape, start the Avid batch digitizing
                    and then go back and continue logging. We tried to keep digitizing
                    non-stop, while Larry and I viewed half-inch dailies in another
                    room on a projection TV. We didn’t actually see the
                    high-def footage projected until we started onlining for previews. Bock: Normally I would be sitting next to
                    the director in a screening room getting some notes and a
                    sense of what he had in mind. We didn’t do that on this
                    show. Jesse looked at dailies on VHS in a trailer. We never
                    looked at the HD material projected and we never sat in a
                    projection room in a group situation. For me, that was very
                    different than previous experiences. How did the experience differ for you creatively?
                   Bock: Creatively, working in hi-def was
                    no different than working on a film show. Were there any problems with the Avid? Things
                    you wish the system would do?  Bock: We knew going into this film that
                    the Avid was the only system that could handle the workload
                    and the amount of temp effects we needed. It was also the
                    only system that would handle the 24P issues. Andersen: We used a Windows NT Avid Symphony,
                    which belonged to the production company, Jersey Films, as
                    well as a Mac-based Meridian Film Composer from Runway, and
                    tied them together with Unity storage. Due to the fact we had a 2 GB project with over 600 files,
                    the system took a long time to auto-save and redraw the bins
                    towards the end of the show. It took an hour to back-up on
                    a DVD-RAM disk each night. I would like to see Avid design
                    an automated backup utility that would work in the background. What were the financial implications of shooting
                    on video?  Andersen: The 24P cameras rent for double the price of a
                    35mm camera, but the cost of HDCAM tape is considerably less
                    than 35mm film, developing and printing. The high definition
                    on-line and color correction bays we used for previews and
                    final cut cost hundreds of dollars per hour, but as more shows
                    shoot in HD, those prices will drop. But even with all the
                    unexpected expenses and the tests we did, the budget for HD
                    post-production was still comparable to a traditional film
                    show.  Did you receive more footage than you would
                    have on a film project?  Bock: Jesse kept the camera rolling to make
                    sure he captured the most comical moments. We received an
                    average of three-and-a-half hours of material a day. By the
                    end of principal photography, we had the equivalent of over
                    750,000 feet of film. Andersen: By the end of the show we had
                    a total of about 400 gigabytes of storage, with our dailies
                    digitized at 14:1. That’s 75 hours of material for a
                    90-minute movie. Did 24P make a difference in how you dealt
                    with visual effects?  Andersen: Larry put me in charge of creating
                    all the temp effects and I tried things that we would never
                    have done had we been working on film. Some of those effects
                    were then recreated during our on-line sessions. For instance,
                    we had a shot of an exploding pigeon where the wire was visible.
                    All we did was paint the wire out, frame by frame. It took
                    20 minutes and probably saved the production a lot of money.
                    In another situation, we needed the character Baby Wipe to
                    walk into a shot. Larry and I looked through all the takes,
                    but the other character in the scene, Baby Powder, had his
                    funniest moments while Baby Wipe was still completely out
                    of frame. Then I suggested that we do a split screen, with
                    Baby Wipe’s half of the frame coming from later in the
                    take than Baby Powder’s side. I did a quick composite
                    on the Avid, and it looked great  How did you online?  Andersen: With the Windows NT Avid Symphony,
                    we could have on-lined the show in NTSC, not hi-def. We would
                    have needed a lot of extra storage to reload the selected
                    material and assemble it uncompressed, so for the projected
                    screenings of the editor’s and director’s cuts
                    we decided to just output the 14:1 compressed image. For our
                    previews, we on-lined using a Fire at Hollywood Digital. The
                    Fire is an uncompressed HD non-linear editing and finishing
                    system made by Discreet. We used it to digitize our HDCAM
                    master tapes, which were then auto- conformed to our EDL,
                    laid back onto HDCAM tape and color-corrected. One big issue
                    with the Fire and with a linear HD online, as well, is that
                    if your effects can’t be described in an EDL, you have
                    to rebuild them in online by hand. After the online, I loaded
                    the finalized Fire effects back into the Avid and cut them
                    in, so that we wouldn’t have to create them again for
                    additional previews. How did you preview?  Andersen: We had three previews, projecting
                    HD using a Panasonic PT 9600U, 1280 x 1024 projector. Bock: Getting ready for previews
                    was very time-consuming, because we had to on-line, color-correct,
                    layback audio to each reel, then tie the film together to
                    make one HD master for screening. But I was surprised by how
                    clean and bright the picture looked when I first saw it projected
                    in high-def. Back when I cut on film, we would preview the
                    print with thousands of splices in it. Then I started cutting
                    on non-linear machines, and the assistants would conform the
                    work print. I thought the film looked a lot better, because
                    it was not being handled as much and had less splices and
                    dirt. Now with high-def, we get to project a spliceless print
                    that has been color-corrected. Although I still prefer the
                    look of a movie shot on film, I was pleasantly surprised that
                    the image didn’t have a pronounced video feel to it.
                    But I was worried about how we would handle the changes from
                    one preview version to the next.  Andersen: The Fire editor recommended archiving
                    our first preview on D5 tape, then loading it back into the
                    Fire. He conformed the first preview to the new cut on his
                    own — we didn’t load any of the onlines into the
                    Avid. We did this for all three previews. The previews gave
                    us our first audience reaction to the 24P image. No one in
                    the audiences noticed that the movie hadn’t been shot
                    on film, and in fact, there were compliments about how nice
                    it looked. Will this film be blown-up to 35mm for release?
                   Andersen: The movie won’t
                    be projected digitally — it’s a film-only release.
                    E-Film did our scan to negative, and Deluxe is doing our prints.
                    Since it cost several dollars per frame to scan the 24P to
                    film, we did a variety of color tests on selected scenes.
                    Our first test was printed on Kodak Vision Print Film 2383.
                    It showed us that we needed a little more contrast in the
                    blacks — the image looked a little flat. For our second
                    test we printed the same negative on 2393 and found that it
                    had more grain and less definition, but the shadows blocked
                    up. So we went back to our color-corrected masters from the
                    first two previews (which had been color-corrected differently),
                    pulled two new scenes, scanned them to 2383 and made a new
                    film print. We discovered we needed to color correct for the
                    film transfer, not for HD projection, and that took care of
                    most of our problems. Each projector is different and we found
                    that, during previews, trying to compensate was pretty frustrating. Bock: Because we were color-correcting on
                    tape, we were able to use power windows, soft clipping, and
                    keying, which wouldn’t be available on film. But I was
                    worried about what would happen to all that work when the
                    show was scanned to film. In general, I was quite surprised
                    at how much it looked like it had been shot on film. Were expectations different from what they
                    would have been on a film project? Were you expected to work
                    faster?  Bock: They always expect you to work faster.
                    I don’t think shooting with the 24P hi-def cameras has
                    changed that.  Overall, was the experience a good one? Would
                    you want to do it again?  Bock: For me, the experience was good. A
                    couple of aspects of the 24P process did create extra work.
                    Jesse asked me to look at everything he shot, including uncircled
                    takes — three to four hours of footage every day, which
                    meant I had less time to edit. I guess I would still prefer
                    to work on projects that are shot on film, but it seems like
                    high-def might become a common medium for features. The answer
                    prints I’ve seen looked good. Maybe not as good as film,
                    but for a comedy like this, I don’t think the audience
                    will ever know how we shot it.  Andersen: With technology advancing so quickly,
                    the image will only get better. I consider myself very lucky
                    to have worked with Larry on the first HD feature for Universal,
                    and I look forward to doing it again. 
 |