Posted on May 12, 2019
I returned Friday from my annual Grand Canyon Raft Trip for Photographers and am playing catch-up on all aspects of my photography life. I’ve barely looked at the my raft trip images, but chose this one for a couple of reasons: first, because I think it perfectly conveys the intimate serenity that always catches me by surprise in this landscape known mostly for it’s broad vistas; and second, because it’s the only image I’ve processed so far.
This is Blacktail Canyon, one of hundreds (thousands?) of narrow slot canyons cutting into the Grand Canyon’s towering walls. Most of them we just float past, sometimes because of the physical challenges required to explore their depths, but usually because there just isn’t time to stop at every slot canyon. On my trips we pick our slots for their photo opportunities, and this year Blacktail Canyon was a particular highlight.
With tall, tightly spaced walls, Blacktail Canyon spends most of its daylight hours in full shade, ideal for photography on sunny days. It doesn’t always have water, but this year’s wet winter meant water in lots of places that don’t always get it. We found the little creek that splits the canyon carrying just enough water to create a series of reflective pools before disappearing into the stream bed, only to reappear farther downstream.
What first drew my eye to this scene was a tiny sapling sprouting from an overhanging ledge, but I soon realized that the tree would best serve me as a visuall element to hold the top of my frame rather than the primary subject. The most interesting thing, I decided, was the blue sky reflection like a jewel embedded in the creek bed.
To create this composition, I dropped my tripod to about a foot above the canyon floor and positioned myself so the lines connecting my primary focal points (the sky reflection, the pair of boulders, and the green tree) created a triangle. Fitting all this into the frame required a vertical orientation of my Sony a7RIII, using virtually the entire width of my Sony 16-35 f/2.8 G lens. Even at this wide focal length, the smooth pebbles at my feet were only about a foot away; getting both the nearby pebbles and glowing (from bounced sunlight) sandstone above the tree sharp, meant choosing my exposure settings and focus point very carefully. My hyperfocal app told me that at f/16, by focusing two feet away, I could achieve my sharpness goal. Watching the rapidly changing sky, I timed my click for the best blend of clouds and sky filling the reflection.
To better understand focus technique, below is an updated version of my Depth of Field article from my Photo Tips section
What’s the point?
It seems like one of photography’s great mysteries is achieving proper focus: the camera settings, where to place the focus point, even the definition of sharpness are all sources of confusion. If you’re a tourist just grabbing snapshots, everything in your frame is likely at infinity and you can just put your camera in full auto mode and click away. But if you’re a photographic artist trying to capture something unique with your mirrorless or DSLR camera and doing your best to have important visual elements objects at different distances throughout your frame, you need to stop letting your camera decide your focus point and exposure settings.
Of course the first creative focus decision is whether you even want the entire frame sharp. While some of my favorite images use selective focus to emphasize one element and blur the rest of the scene, most (but not all) of what I’ll say here is about using hyperfocal techniques to maximize depth of field (DOF). I cover creative selective focus in much greater detail in another Photo Tip article: Creative Selective Focus.
Beware the “expert”
I’m afraid that there’s some bad, albeit well-intended, advice out there that yields just enough success to deceive people into thinking they’ve got focus nailed, a misperception that often doesn’t manifest until an important shot is lost. I’m referring to the myth that you should focus 1/3 of the way into the scene, or 1/3 of the way into the frame (two very different things, each with its own set of problems).
For beginners, or photographers whose scene doesn’t include subjects from near to far, the 1/3 technique may be a useful rule of thumb. But taking the 1/3 approach to focus requires that you understand DOF and the art of focusing well enough to know when 1/3 won’t work, and how to adjust your focus point and settings. And once you achieve that level of understanding, you may as well do it the right way from the start. Focus control becomes especially important in those scenes where missing the focus point by just a few feet or even inches can make or break and image.
Back to the basics
Understanding a few basic focus truths will help you make focus decisions:
- A lens’s aperture is the opening that allows light to reach your sensor—the bigger this opening, the more light gets in, but also the smaller your DOF.
- Aperture is measured in f-stops, which is the lens’s focal length divided by the aperture’s diameter; the higher the f-number, the smaller the aperture and the greater the DOF. So f/8 is actually a bigger aperture (with less DOF) than f/11. This understanding becomes second nature, but if you’re just learning it’s helpful to think of f/stops this way: The higher the f-number, the greater the depth of field. Though they’re not exactly the same thing, photographers usually use f-stop and aperture interchangeably.
- Regardless of its current f-stop setting, a DSLR camera maximizes the light in its viewfinder by always showing you the scene at the lens’s widest aperture. All this extra light makes it easier to compose and focus, but unless your exposure is set for the widest aperture (which it shouldn’t be unless you have a very specific reason to limit your depth of field or maximize light), the image you capture will have more DOF than you see in the viewfinder. The consequence is that you usually can’t see how much of your scene is in focus when you compose. Most cameras have a DOF preview button that temporarily closes the lens down to the f-stop you have set—this shows the scene at its actual DOF, but also darkens the viewfinder considerably (depending on how small your aperture is), making it far more difficult to see the scene.
- For any focus point, there’s only one (infinitely thin) plane of perfect sharpness, regardless of the focal length and f-stop—everything in front of and behind the plane containing your focus point (and parallel to the sensor) will be some degree of less than maximum sharpness. As long as the zone of less than perfect sharpness isn’t perceptible, it’s considered “acceptably sharp.” When that zone becomes visible, that portion of the image is officially “soft.” Acceptable sharpness varies with the display size and viewing distance.
- The zone of acceptable sharpness extends a greater distance beyond the focus point than it does in front of the focus point. If you focus on that rock ten feet in front of you, rocks three feet in front of you may be out of focus, but a tree fifty feet away could be sharp. I’ll explain more about this later.
- While shorter focal lengths may appear to provide more depth of field, believe it or not, DOF doesn’t actually change with focal length. What does change is the size of everything in the image, so as your focal length increases, your apparent DOF decreases. So you really aren’t gaining more absolute DOF with a shorter focal length, the softness just won’t be as visible. When photographers talk about DOF, they’re virtually always talking about apparent DOF—the way the image looks. (That’s the DOF definition I use here too.)
- The closer your focus point, the narrower your DOF (range of front-to-back sharpness). If you focus your 24mm lens on a butterfly sunning on a poppy six inches from your lens, your DOF is so narrow that it’s possible parts of the poppy will be out of focus; if you focus the same lens on a tree 100 feet away, the mountains behind the tree are sharp too.
Depth of field discussions are complicated by the fact that “sharp” is a moving target that varies with display size and viewing distance. But it’s safe to say that all things equal, the larger your ultimate output and closer the intended viewing distance, the more detail your original capture should contain.
To capture detail a lens focuses light on the sensor’s photosites. Remember using a magnifying glass to focus sunlight and ignite a leaf when you were a kid? The smaller (more concentrated) the point of sunlight, the sooner the smoke appeared. In a camera, the finer (smaller) a lens focuses light on each photosite, the more detail the image will contain at that location. So when we focus we’re trying to make the light striking each photosite as concentrated as possible.
In photography we call that small circle of light your lens makes for each photosite its “circle of confusion.” The larger the CoC, the less concentrated the light and the more blurred the image will appear. Of course if the CoC is too small to be seen as soft, either because the print is too small or the viewer is too far away, it really doesn’t matter. In other words, areas of an image with a large CoC (relatively soft) can still appear sharp if small enough or viewed from far enough away. That’s why sharpness can never be an absolute term, and we talk instead about acceptable sharpness that’s based on print size and viewing distance. It’s actually possible for the same image to be sharp for one use, but too soft for another.
So how much detail do you need? The threshold for acceptable sharpness is pretty low for an image that just ends up on an iPhone or an 8×10 calendar on the kitchen wall, but if you want that image to fill the wall above the sofa, acceptable sharpness requires much more detail. And as your print size increases (and/or viewing distance decreases), the CoC that delivers acceptable sharpness shrinks correspondingly.
Many factors determine the a camera’s ability to record detail. Sensor resolution of course—the more resolution your sensor has, the more important it becomes that to have a lens that can take advantage of that extra resolution. And the more detail you want to capture with that high resolution sensor and tack-sharp lens, the more important your depth of field and focus point decisions become.
The foundation of a sound approach to maximizing sharpness for a given viewing distance and image size is hyperfocal focusing, an approach that uses viewing distance, f-stop, focal length, and focus point to ensure acceptable sharpness.
The hyperfocal point is the focus point that provides the maximum depth of field for a given combination of sensor size, f/stop, and focal length. Another way to express it is that the hyperfocal point is the closest you can focus and still be acceptably sharp to infinity. When focused at the hyperfocal point, your scene will be acceptably sharp from halfway between your lens and focus point all the way to infinity. For example, if the hyperfocal point for your sensor (full frame, APS-C, 4/3, or whatever), focal length, and f-stop combinition is twelve feet away, focusing there will give you acceptable sharpness from six feet (half of twelve) to infinity—focusing closer will soften the distant scene; focusing farther will keep you sharp to infinity but extend the area of foreground softness.
Because the hyperfocal variable (sensor size, focal length, f-stop) combinations are too numerous to memorize, we usually refer to an external aid. That used to be awkward printed tables with long columns and rows displayed in microscopic print, the more precise the data, the smaller the print. Fortunately, those have been replaced by smartphone apps with more precise information in a much more accessible and readable form. We plug in all the variables and out pops the hyperfocal point distance and other useful information
It usually goes something like this:
- Identify the composition
- Determine the closest thing that must be sharp (right now I’m assuming you want sharpness to infinity)
- Dig the smartphone from one of the 10,000 pockets it could be in
- Open the hyperfocal app and plug in the sensor size (usually previously set by you as the default), f-stop, and a focus distance
- Up pops the hyperfocal distance (and usually other info of varying value)
You’re not as sharp as you think
Since people’s eyes start to glaze over when CoC comes up, they tend to use the default returned by the smartphone app. But just because the app tells you you’ve nailed focus, don’t assume that your work is done. An often overlooked aspect of hyperfocal focusing is that app makes assumptions that aren’t necessarily right, and in fact are probably wrong.
The CoC your app uses to determine acceptable sharpness is a function of sensor size, display size, and viewing distance. But most app’s hyperfocal tables assume that you’re creating an 8×10 print that will be viewed from a foot away—maybe valid 40 years ago, but not in this day of mega-prints. The result is a CoC three times larger than the eye’s ability to resolve.
That doesn’t invalidate hyperfocal focusing, but if you use published hyperfocal data from an app or table, your images’ DOF might not be as ideal as you think it is for your use. If you can’t specify a smaller CoC in your app, I suggest that you stop-down a stop or so more than the app/table indicates. On the other hand, stopping down to increase sharpness is an effort of diminishing returns, because diffraction increases as the aperture shrinks and eventually will soften the entire image—I try not to go more than a stop smaller than my data suggests.
Keeping it simple
As helpful as a hyperfocal app can be, whipping out a smartphone for instant in-the-field access to data is not really conducive to the creative process. I’m a big advocate of keeping photography as simple as possible, so while I’m a hyperfocal focus advocate in spirit, I don’t usually use hyperfocal data in the field. Instead I apply hyperfocal principles in the field whenever I think the margin of error gives me sufficient wiggle room.
Though I don’t often use the specific hyperfocal data in the field, I find it helps a lot to refer to hyperfocal tables when I’m sitting around with nothing to do. So if I find myself standing in line at the DMV, or sitting in a theater waiting for a movie (I’m a great date), I open my iPhone hyperfocal app and plug in random values just to get a sense of the DOF for a given f-stop and focal length combination. I may not remember the exact numbers later, but enough of the information sinks in that I accumulate a general sense of the hyperfocal DOF/camera-setting relationships.
Finally, something to do
Unless I think I have very little DOF margin for error in my composition, I rarely open my hyperfocal app in the field. Instead, once my composition is worked out and have determined the closest object I want sharp—the closest object with visual interest (shape, color, texture), regardless of whether it’s a primary subject.
- If I want to be sharp to infinity and my closest foreground object (that needs to be sharp) is close enough to hit by tossing my hat, I need a fair amount of DOF. If my focal length is pretty wide, I might skip the hyperfocal app, stop down to f/16, and focus a little behind my foreground object. But if I’m at a fairly long focal length, or my closest object is within arm’s reach, I have very little margin for error and will almost certainly refer to my hyperfocal app.
- If I could hit my foreground object with a baseball and my focal length is 50mm (or so) or less, I’ll probably go with f/11 and just focus on my foreground object. But as my focal length increases, so does the likelihood that I’ll need to refer to my hyperfocal app.
- If it would take a gun to reach my closest object (picture a distant peak), I choose an f-stop between f/8 and f/11 and focus anywhere in the distance.
Of course these distances are very subjective and will vary with your focal length and composition (not to mention the strength of your pitching arm), but you get the idea. If you find yourself in a small margin for error focus situation without a hyperfocal app (or you just don’t want to take the time to use one), the single most important thing to remember is to focus behind your closest subject. Because you always have sharpness in front of your focus point, focusing on the closest subject gives you unnecessary near sharpness at the expense of distant sharpness. By focusing a little behind your closest subject, you’re increasing the depth of your distant sharpness while (if you’re careful) keeping your foreground subject within the zone of sharpness in front of the focus point.
And finally, foreground softness, no matter how slight, is almost always a greater distraction than slight background softness. So, if it’s impossible to get all of your frame sharp, it’s usually best to ensure that the foreground is sharp.
Why not just automatically set the aperture to f/22 and be done with it? I thought you’d never ask. Without delving too far into the physics of light and optics, let’s just say that there’s a not so little light-bending problem called “diffraction” that robs your images of sharpness as your aperture shrinks—the smaller the aperture, the greater the diffraction. Then why not choose f/2.8 when everything’s at infinity? Because lenses tend to lose sharpness at their aperture extremes, and are generally sharper in their mid-range f-stops. So while diffraction and lens softness don’t sway me from choosing the f-stop that gives the DOF I want, I try to never choose an aperture bigger or smaller than I need.
Now that we’ve let the composition determine our f-stop, it’s (finally) time to actually choose the focus point. Believe it or not, with this foundation of understanding we just established, focus becomes pretty simple. Whenever possible, I try to have elements throughout my frame, often starting near my feet and extending far into the distance. When that’s the case I stop down focus on an object slightly behind my closest subject (the more distant my closest subject, the farther behind it I can focus).
When I’m not sure, or if I don’t think I can get the entire scene sharp, I err on the side of closer focus to ensure that the foreground is sharp. Sometimes before shooting I check my DOF with the DOF preview button, allowing time for my eye to adjust to the limited light. And when maximum DOF is essential and I know my margin for error is small, I don’t hesitate to refer to the DOF app on my iPhone.
A great thing about digital capture is the instant validation of the LCD—when I’m not sure, or when getting it perfect is absolutely essential, after capture I pop my image up on the LCD, magnify it to maximum, check the point or points that must be sharp, and adjust if necessary. Using this immediate feedback to make instant corrections really speeds the learning process.
Sometimes less is more
The depth of field you choose is your creative choice, and no law says you must maximize it. Use your camera’s limited depth of field to minimize or eliminate distractions, create a blur of background color, or simply to guide your viewer’s eye. Focusing on a near subject while letting the background go soft clearly communicates the primary subject while retaining enough background detail to establish context. And an extremely narrow depth of field can turn distant flowers or sky into a colorful canvas for your subject.
There’s no substitute for experience
No two photographers do everything exactly alike. Determining the DOF a composition requires, the f-stop and focal length that achieves the desired DOF, and where to place the point of maximum focus, are all part of the creative process that should never be left up to the camera. The sooner you grasp the underlying principles of DOF and focus, the sooner you’ll feel comfortable taking control and conveying your own unique vision.
About this image
Yosemite may not be New England, but it can still put on a pretty good fall color display. A few years ago I arrived at Valley View on the west side of Yosemite Valley just about the time the fall color was peaking. I found the Merced River filled with reflections of El Capitan and Cathedral Rocks, framed by an accumulation of recently fallen leaves still rich with vivid fall color.
To emphasize the colorful foreground, I dropped my tripod low and framed up a vertical composition. I knew my hyperfocal distance at 24mm and f/11 would be 5 or 6 feet, but with the scene ranging from the closest leaves at about 3 feet away out to El Capitan at infinity, I also knew I’d need to be careful with my focus choices. For a little more margin for error I stopped down to f/16, then focused on the nearest rocks which were a little less than 6 feet away. As I usually do when I don’t have a lot of focus wiggle room, I magnified the resulting image on my LCD and moved the view from the foreground to the background to verify front-to-back sharpness.
Playing with Depth: A Gallery of Focus
Click an image for a closer look and slide show. Refresh the screen to reorder the display.
Posted on April 28, 2019
Who doesn’t like blue skies? Well…, me. I’ll say it again: I. Don’t. Like. Blue. Skies. At least not for photography. As pleasant as blue sky is for a stroll on the beach or picnic at the park, it’s just plain boring in a picture, and I do everything in my power to avoid an empty sky in my images. Don’t believe me? Check out my Master Gallery, count the number of images that have sky that occupies more than 1/5 of my frame with nothing else in it, then report back to me.
Fortunately, there are lots of antidotes for the blues: clouds, stars, a rainbow, a sunstar, and the moon can all redeem a blue sky. And sometimes the best cure is simply no sky at all.
A little planning
Whether it’s the Milky Way from the bottom of the Grand Canyon, or lightning on the top; a full moon at sunrise in Death Valley, or at sunset in Yosemite, most of my workshops and personal photo trips are timed to coincide with something interesting happening in the sky.
Some helpful links (from my Photo Tips section)
This year’s Yosemite spring workshops were scheduled to coincide with the full moon. In these workshops the priority was a “moonbow” (lunar rainbow) on Yosemite Fall, but we also had opportunities to photograph a couple of moonrises in each workshop. By spring, as the sun moves north in our sky, the full moon (which rises opposite the setting sun) has shifted south and out of alignment with Half Dome from my favorite Yosemite Valley locations. So I seek other Yosemite features to align with the rising moon.
The Bridalveil Fall view from Big Oak Flat Road is a particular spring moonrise favorite. Not just for the way it aligns the moon and Bridalveil Fall, but also because it gives the option of focal lengths from wide to long. I’d planned a rising full moon at this spot for my April’s workshop all the way back when I scheduled it over a year ago, but as clouds drifted in (be careful what you wish for), I began to have doubts that we’d see the moon at all.
Waiting there it was hard to tell if the clouds were thick enough to completely obscure the moon. As you can see from this image, they weren’t, though they certainly did soften the lunar features. As with most of my moonrise shoots, I started long when the moon crested the horizon, then gradually widened as the moon rose. The clouds turned out to be a blessing; not only were they thin enough to enable the moon shine through, they provided the ideal canvas for a vivid sunset display.
A few tips for avoiding the blues
- If the sky is boring, eliminate or minimize it
- What’s happening in the clouds can be as important as what’s happening in the landscape—give the sky as much attention as the sky as you make your composition
- The moon can carry a significant amount of empty sky—even if it’s a small part of an otherwise blank sky, the moon makes a wonderful accent to any scene
- What’s more interesting, the sky or the landscape? Determine the position of your horizon in your frame by the relative visual interest between the sky and landscape—the better the sky, the more of your frame it should occupy, and vice-versa. And if the sky and landscape are equally interest (beautiful), there’s nothing wrong with splitting the frame down the middle
Look to the Sky
Posted on April 21, 2019
There are many (many!) beautiful sights in Yosemite, but when most people think about Yosemite, they think about waterfalls and granite. The granite is forever (virtually), but Yosemite’s waterfalls come and go with the season: exploding from the granite walls in spring, most of Yosemite’s waterfalls are bone dry by summer’s end. And some years are better than others—three springs ago, Bridalveil and Yosemite Falls were barely a trickle, too dry to photograph (unprecedented in my lifetime). The next spring the deafening roar of waterfalls was back, echoing throughout Yosemite Valley.
Moonbow, April 18, 2019
I just returned from my annual Yosemite Moonbow and Dogwood photo workshop on Friday night (technically, it was early Saturday morning). The dogwood are just starting to pop, but the waterfalls are going strong, with enough snow in the high Sierra bank to keep them roaring through summer.
My group photographed more waterfall rainbows than I could count, on both Bridalveil and Yosemite Falls, but the highlight was Thursday night’s lunar rainbow (moonbow) shoot on the bridge beneath Lower Yosemite Fall. Nothing compares to the first time seeing a moonbow. A shimmering silver arc, a moonbow is clearly visible to the naked eye—proper exposure in a camera reveals the moonbow’s vivid color.
A “practice” moonlight shoot the previous night helped prepare everyone for the difficulties of photographing in the dark. And while my group came prepared for moonlight photography, the crowds and mist make things difficult even for the seasoned veteran. The crowds weren’t too bad this year, but while lots of water in the fall means a better moonbow, it also means a wetter photographer.
I feared that the thin cloud cover that had delivered a spectacular sunset just as the full moon rose just an hour or so earlier, would douse the moonlight necessary for a moonbow, but that turned out to be a non-factor. One problem was contrails, more than I’ve ever seen. Some chose to leave the sky (or most of the sky) out of their frame; I opted to include the sky, then carefully execute a contrailecotmy in Photoshop.
Because most of my time on the bridge is spent assisting the group, I only got to click a handful of frames. I started on the (drier) paved open area before the bridge, but after working with a workshop participant on the bridge, I decided the view there was worth getting wet.
I went wider with this year’s images than previous years, using my Sony 12-24 G lens on my Sony a7RIII camera. I focused on the moon, then turned around and set up my composition. Concerned about too much water on my front lens element, I bumped my ISO to 1600 to keep my shutter speed at 10 seconds or faster. When I was ready to click, I wiped down the front of my lens with a towel that I lifted just as my shutter clicked.
Yosemite April 2019 Images
(processed so far)
I just scheduled my 2020 Yosemite Spring photo workshops, April 5-8 and May 4-7. Both are timed for the full moon to maximize our moonbow chances. And of course it’s not all about waterfalls and rainbows—this year’s spring workshops included some spectacular clearing storms, beautiful moonrises, and brilliant poppies. In addition to great photography, you’ll improve your photo skills with daily training and image reviews. You’ll also have lots of fun.
Click an image for a closer look and to view slide show.
Posted on April 17, 2019
The annual Grand Canyon monsoon is known for its spectacular electrical storms, but let’s not forget the rainbows that often punctuate these storms. A rainbow requires rain, sunlight, and the right viewing angle—given the ephemeral nature of a monsoon thunderstorm, it’s usually safe to assume that the sun probably isn’t far behind. To experience a rainbow after a Grand Canyon monsoon storm, all it takes is some basic knowledge, a little faith, and some good fortune.
To help with the knowledge part, I’m sharing the how-and-why of rainbows, excerpted from my just updated Rainbow article in my Photo Tips section. For the faith and good fortune part, read “The story of this image” at the bottom of this post.
Most people understand that a rainbow is light spread into various colors by airborne water drops. Though a rainbow can seem like a random, unpredictable phenomenon, the natural laws governing rainbow are actually quite specific and predictable, and understanding these laws can help photographers anticipate a rainbow and enhance its capture.
Let there be light
Energy generated by the sun bathes Earth in continuous electromagnetic radiation, its wavelengths ranging from extremely short to extremely long (and every wavelength in between). Among the broad spectrum of electromagnetic solar energy we receive are ultra-violet rays that burn our skin, infrared waves that warm our atmosphere, and a very narrow range of wavelengths the human eye sees.
These visible wavelengths are captured by our eyes and interpreted by our brain. When our eyes take in light comprised of the full range of visible wavelengths, we perceive it as white (colorless) light. Color registers when some wavelengths are more prevalent than others. For example, when light strikes an opaque (solid) object such as a tree or rock, some of its wavelengths are absorbed; the wavelengths not absorbed are scattered (reflected). Our eyes capture this scattered light, send the information to our brains, which interprets it as a color. When light strikes water, some is absorbed, some passes through to reveal the submerged world, and some light is reflected by the surface as a reflection.
To understand the interaction of water and light that creates a rainbow, it’s simplest to visualize what happens when sunlight strikes a single drop. Light entering a water drop refracts (bends), with different wavelengths refracting different amounts, which separates the originally homogeneous white light into the myriad colors of the spectrum.
But simply separating the light into its component colors isn’t enough to create a rainbow—if it were, we’d see a rainbow whenever light strikes water. Seeing the rainbow spectrum caused by refracted light requires that the refracted light be returned to our eyes somehow.
A raindrop isn’t flat like a sheet of paper, it’s spherical, like a ball. Light that was refracted (and separated into multiple colors) as it entered the front of the raindrop, continues through to the back of the raindrop, where some is reflected. Red light reflects back at about 42 degrees, violet light reflects back at about 40 degrees, and the other spectral colors reflect back between 42 and 40 degrees. What we perceive as a rainbow is this reflection of the refracted light—notice how the top color of the primary rainbow is always red, the longest visible wavelength; the bottom color is always violet, the shortest visible wavelength.
Follow your shadow
Every raindrop struck by sunlight creates a rainbow. But just as the reflection of a mountain peak on the surface of a lake is visible only when viewed from the angle the reflection bounces off the lake’s surface, a rainbow is visible only when you’re aligned with the 40-42 degree angle at which the raindrop reflects the spectrum of rainbow colors.
Fortunately, viewing a rainbow requires no knowledge of advanced geometry. To locate or anticipate a rainbow, picture an imaginary straight line originating at the sun, entering the back of your head, exiting between your eyes, and continuing down into the landscape in front of you—this line points to the “anti-solar point,” an imaginary point exactly opposite the sun. With no interference, a rainbow would form a complete circle, skewed 42 degrees from the line connecting the sun and the anti-solar point—with you at the center. (We don’t see the entire circle because the horizon usually gets in the way.)
Because the anti-solar point is always at the center of the rainbow’s arc, a rainbow will always appear exactly opposite the sun (the sun will always be at your back). It helps to remember that your shadow always points toward the anti-solar point. So when you find yourself in direct sunlight and rain, locating a rainbow is as simple as following your shadow and looking skyward—if there’s no rainbow, the sun’s probably too high.
High or low
Sometimes a rainbow appears as a majestic half-circle, arcing high above the distant terrain; other times it’s merely a small circle segment hugging the horizon. As with the direction of the rainbow, there’s nothing mysterious about its varying height. Remember, every rainbow would form a full circle if the horizon didn’t get in the way, so the amount of the rainbow’s circle you see (and therefore its height) depends on where the rainbow’s arc intersects the horizon.
While the center of the rainbow is always in the direction of the anti-solar point, the height of the rainbow is determined by the height of the anti-solar point, which will always be exactly the same number of degrees below the horizon as the sun is above the horizon. It helps to imagine the line connecting the sun and the anti-solar point as a fulcrum, with you as the pivot—picture yourself in the center of a teeter-totter: as one seat rises above you, the other drops below you. That means the lower the sun, the more of its circle you see and the higher it appears above the horizon; conversely, the higher the sun, the less of its circle is above the horizon and the flatter (and lower) the rainbow will appear.
Assuming a flat, unobstructed scene (such as the ocean), when the sun is on the horizon, so is the anti-solar point (in the opposite direction), and half of the rainbow’s 360 degree circumference will be visible. But as the sun rises, the anti-solar point drops—when the sun is more than 42 degrees above the horizon, the anti-solar point is more than 42 degrees below the horizon, and the only way you’ll see a rainbow is from a perspective above the surrounding landscape (such as on a mountaintop or on a canyon rim).
Of course landscapes are rarely flat. Viewing a scene from above, such as from atop Mauna Kea or from the rim of the Grand Canyon, can reveal more than half of the rainbow’s circle. From an airplane, with the sun directly overhead, all of the rainbow’s circle can be seen, with the plane’s shadow in the middle.
Double Your pleasure
Not all of the light careening about a raindrop goes into forming the primary rainbow. Some of the light slips out the back of the raindrop to illuminate the sky, and some is reflected inside the raindrop a second time. The refracted light that reflects a second time before exiting creates a secondary, fainter rainbow skewed 50 degrees from the anti-solar point. Since this is a reflection, the colors of the secondary rainbow are reversed from the primary rainbow.
And if the sky between the primary and secondary rainbows appears darker than the surrounding sky, you’ve found “Alexander’s band.” It’s caused by all the light machinations I just described—instead of all the sunlight simply passing through the raindrops to illuminate the sky, some of the light was intercepted, refracted, and reflected by the raindrops to form our two rainbows, leaving less light for the sky between the rainbows.
Waterfalls are easy
Understanding the optics of a rainbow has practical applications for photographers. Not only does it help you anticipate a rainbow before it happens, it also enables you to find rainbows in waterfalls.
Unlike a rainbow caused by rain, which requires you to be in exactly the right position to capture the incongruous convergence of rainfall and sunshine, a waterfall rainbow can be predicted with clock-like precision—just add sunshine.
Yosemite is my location of choice, but there’s probably a waterfall or two near you that will deliver. Just figure out when the waterfall gets direct sunlight early or late in the day, then put yourself somewhere on the line connecting the sun and the waterfall. And if you have an elevated vantage point, you’ll find that the sun doesn’t even need to be that low in the sky.
Understanding rainbow optics can even help you locate rainbows that aren’t even visible to the naked eye. A “moonbow” (lunar rainbow) is a rarely witnessed and wonderful phenomenon that follows all the natural rules of a daylight rainbow. But instead of resulting from direct sunlight, a moonbow is caused by sunlight reflected by the moon.
Moonlight isn’t bright enough to fully engage the cones in your eyes that reveal color, though in bright moonlight you can see the moonbow as an arcing monochrome band. But a camera on a sturdy tripod can use its virtually unlimited shutter duration to accumulate enough light to bring out a moonbow in full living color. Armed with this knowledge, all you need to do is put yourself in the right location at the right time.
The story of this image
Following a nice sunrise at the always beautiful Point Imperial, the Grand Canyon Monsoon photo workshop group spent two hours near Bright Angel Point photographing a spectacular electrical storm that delivered multiple lightning captures to everyone in the group. When the storm moved too close and drove us to safety (we’re resilient and adventuresome, not stupid), it would have been easy call it a day and tally our bounty. I mean, who likes getting rained on? Photographers, that’s who.
Don Smith and I herded our group into the cars and headed to Cape Royal Road, where we could follow the Grand Canyon’s East Rim above Marble Canyon all the way to Cape Royal. Knowing that monsoon showers are fairly localized, the plan was to drive out of the cell that was dumping on us at the lodge and either shoot back at it, or (more likely) find another cell firing out over the canyon. In the back of my mind though was the hope for a rainbow above the canyon—dropping in the west, the sun was perfectly positioned for rainbows in the east.
The rainbow appeared just after we passed the Point Imperial Road junction, arcing high above the forest. Climbing through the trees toward the rim (and its views of Marble Canyon), my urgency intensified with the rainbow’s vivid color, but we were stuck behind a meandering tourist who clearly had different priorities. As tempted as I was to pass him, I knew that would be a mistake with three more cars following me. So we poked along at a glacial pace. After what felt like hours, screeched to a halt at the Vista Encantada parking area with the rainbow hanging in there—I swear everyone was out of the car and scrambling for their gear before I came to a complete stop.
With a full rainbow above an expansive view, I opted for my Sony 12-24 lens on my a7RII, but immediately began to question that choice. While Vista Encantada offers a very pretty view, it’s not my favorite scene to photograph because of the less-than-photogenic shrubbery in the foreground—a telephoto lens definitely would have worked better to eliminate the foreground, but I wanted more rainbow. So after a few failed attempts to find a composition at the conventional vista, I sprinted into the woods to find something better. This turned out to be a wise choice, as the shrubs here were replaced with (much more photogenic) mature evergreens.
In a perfect world I’d have found an unobstructed view into the Grand Canyon, but as photographers know, the world is rarely perfect. Committed to my wide lens, I decided to use the nearby evergreens as my foreground, moving back just far enough for the rainbow to clear their crowns. Composing wide enough to include the trees top-to-bottom also allowed me to include all of the rainbow—suddenly my 12-24 lens choice was genius!
After finishing at Vista Encantada we continued down the road and photographed another rainbow from Roosevelt Point, then wrapped up the day with a sunset for the ages at Cape Royal. A great day indeed, all thanks to monsoon weather that would have kept most tourists indoors.
A Gallery of Rainbows
Click an image for a closer look and to view slide show.
Posted on April 14, 2019
I’m often asked if I placed a leaf, moved a rock, or “Photoshopped” a moon into an image. Usually the tone is friendly curiosity, but sometimes it’s tinged with hints of suspicion that can border on accusation. While these questions are an inevitable part of being a photographer today, I suspect that I get more than my share because I aggressively seek out naturally occurring subjects to isolate and emphasize in my frame. But regardless of the questioner’s tone, my answer is always a cheerful and unapologetic, “No.”
We all know photographers who have no qualms about staging their scenes to suit their personal aesthetics. The rights and wrongs of that are an ongoing debate I won’t get into, other than to say that I have no problem when photographers arrange their scenes openly, with no intent to deceive. But photography must be a source of pleasure, and my own photographic pleasure derives from discovering and revealing nature, not manufacturing it. I don’t like arranging scenes because I have no illusions that I can improve nature’s order, and am confident that there’s enough naturally occurring beauty to keep me occupied for the rest of my life.
Order vs. chaos
As far as I’m concerned, nature is inherently ordered. In fact, in the grand scheme, “nature” and “order” are synonyms. But humans go to such lengths to control, contain, and manage the natural world that we’ve created a label for our failure to control nature: Chaos. Despite its negative connotation, what humans perceive as “chaos” is actually just a manifestation of the universe’s inexorable push toward natural order.
Let’s take a trip
Imagine all humans leave Earth for a scenic tour of the Milky Way. While we’re gone, no lawns are mowed, no buildings maintained, no fires extinguished, no floods controlled, no Starbucks built. Let’s say we return in 100 Earth years*. While the state of things would no doubt be perceived as chaotic, the reality is that our planet would in fact be closer to its natural state. And the longer we’re away, the more human-imposed “order” would be replaced by natural order.
* Since this is my fantasy, I’ve chartered a spaceship that accommodates all of humankind and travels at 90 percent of the speed of light. While Earth has indeed aged 100 years during our holiday, we travelers return only a year older. (Dubious? Don’t take my word for it, ask Albert Einstein.)
What does all this have to do with raindrops on a poppy?
Read the story of this saturated shoot in my All Wet blog post
Venturing outdoors with a camera and the mindset that nature is inherently ordered makes me feel like a treasure hunter—I know the treasure is there, I just have to find it. Patterns and relationships hidden by human interference and the din of 360 degree multi-sensory input, further obscured by human bias, snap into coherence when I find the right perspective.
Finding water droplets to photograph can be as simple as picking a subject and squirting it with a spray bottle of water or (better still) glycerin. But what fun is that? If I’d have been staging this, I probably would have insisted on an open poppy, maybe with more and bigger drops. But that’s not what Nature gave me this soggy afternoon. So I photographed this raindrop festooned poppy (and many others) the old fashioned way—within minutes I was as wet as the poppy, and (to quote the immortal Cosmo Kramer) lovin’ every minute of it.
Finding Natural Order
Click an image for a closer look and to view slide show.
Posted on April 7, 2019
Every year for the last 10 (or so) years I’ve traveled to the Grand Canyon during the Southwest summer monsoon to photograph lightning. Not only have I captured hundreds of lightning strikes and lived to tell about it (yay), I’ve learned a lot. A couple of years ago I added an article sharing my insights on photographing lightning to my photo tips section. With lightning season upon (or almost upon) us here in the United States, I’ve updated my article with new images and additional info. You can still find the article (with updates) in my Photo Tips section, but I’m re-posting it here in my regular blog feed as well.
Read the story of this image at the bottom of this post, just above the gallery of lightning images.
How to Photograph Daylight Lightning Without Getting Killed (Probably)
Let’s start with the given that lightning is dangerous, and if “safety first” is a criterion for intelligence, photographers are stupid. So combining photographers and lightning is a recipe for disaster.
Okay, seriously, because lightning is both dangerous and unpredictable, before attempting anything that requires you to be outside during an electrical storm, it behooves you to do your homework. And the more you understand lightning, how to avoid it and stay safe in its presence, the greater your odds of living to take more pictures. Not only will understanding lightning improve your safety, a healthy respect for lightning’s fickle power will also help you anticipate and photograph lightning.
Lightning is an electrostatic discharge that equalizes the negative/positive polarization between two objects. In fact, when you get shocked touching a doorknob, you’ve been struck by lightning. The cause of polarization during electrical storms isn’t completely understood, but it’s generally accepted that the extreme vertical convective air motion (convection is up/down circular flow caused when less-dense warm air rises, becomes more dense as it cools with elevation, and ultimately becomes cool/dense enough to fall. Convection is also what causes bubbling in boiling water. Convection in a thunderstorm carries positively charged molecules upward and negatively charged molecules downward. Because opposite charges attract each other, the extreme polarization (positive charge at the top of the cloud, negative charge near the ground) is quickly (and violently) equalized: lightning.
With lightning comes thunder, the sound of air expanding explosively when heated by a 50,000 degree jolt of electricy. The visual component of the lightning bolt that caused the thunder travels to you at the speed of light, over 186,000 miles per second (virtually instantaneous regardless of your distance on Earth). But lightning’s aural component, thunder, only travels at the speed of sound, a little more than 750 miles per hour—a million times slower than light. Knowing that the thunder occurred at the same time as the lightning flash, and how fast both travel, we can compute the approximate distance of the lightning strike. At 750 miles per hour, thunder will travel about a mile in about five seconds: Dividing the time between the lightning’s flash and the thunder’s crash by five gives you the lightning’s distance in miles; divide the interval by three for the distance in kilometers. If five seconds pass between the lightning and the thunder, the lightning struck about one mile away; fifteen seconds elapsed means it’s about three miles away.
The 30 (or so) people killed by lightning in the United States each year had one thing in common with the rest of us: they didn’t believe they’d be struck by lightning when they started whatever it was they were doing when they were struck. The only sure way to be safe in an electrical storm is to be in a fully enclosed structure or metal-framed vehicle, away from open windows, plumbing, wiring, and electronics.
While there’s no completely safe way to photograph lightning, it doesn’t hurt to improve your odds of surviving to enjoy the fruits of your labor. (Unfortunately, photographing lightning usually requires being outside.) Most lightning strikes within a six mile radius of the previous strike. So, if less than thirty seconds elapses between the flash and bang, you’re too close. And since “most” doesn’t mean “all,” it’s even better to allow a little margin for error. Thunder isn’t usually audible beyond ten miles—if you can hear the thunder, it’s safe to assume that you’re in lightning range.
But if you absolutely, positively must be outside with the lightning crashing about you, or you simply find yourself caught outside with no available shelter, there are few things you can do to reduce the chance you’ll be struck:
- Avoid water
- Avoid high ground
- Avoid exposed areas
- Avoid metal or electronic objects
- Avoid tall objects such as trees and open structures (and tripods)
- Stay at least fifteen feet from other people
- Do not lie down
- If you’re surrounded by trees, position yourself near shorter trees, as far from trunks as possible
- Crouch with your feet together and your hands covering your ears
- A lightning strike is often preceded by static electricity that makes your hair stand on end and an ozone smell (best described as the smell of electricity—I think of bumper cars at the amusement park, or the smell of my electric slot cars when I was a kid)—if your hair starts to stand up and/or you notice a distinct odor that could be ozone, follow as many of the above steps as you can, as quickly as possible (often you’ll only have time to crouch)
Photographing lightning at night is mostly a matter of pointing your camera in the right direction with a multi-second shutter speed and hoping the lightning fires while your shutter’s open—pretty straightforward. Photographing daylight lightning is a little more problematic. It’s usually over before you can react, so without a lightning sensor to recognize lightning and click your shutter, success is largely dumb luck (few people are quick enough see it and click). And using a neutral density filter to stretch the exposure time out to 20 or 30 seconds sounds great in theory, but a lightning bolt with a life measured in milliseconds, captured in an exposure measured in multiple seconds, will almost certainly lack the contrast necessary to be be even slightly visible.
Lightning Trigger: The best tool for the job
Most lightning sensors (all?) attach to your camera’s hot shoe and connect via a special cable to the camera’s remote-release port. When engaged, the sensor fires the shutter (virtually) immediately upon detecting lightning, whether or not the lightning is visible to the eye or camera. With many lightning sensors from which to choose, before I bought my first one I did lots of research. I ended up choosing the sensor that was the consensus choice among photographers I know and trust: Lightning Trigger from Stepping Stone Products in Dolores, CO. At around $350 (including the cable), the Lightning Trigger is not the cheapest option, but after many leading lightning-oriented photo workshops, I can say with lots of confidence that lightning sensors are not generic products, and the internal technology matters a lot. Base on my own results and observations, the Lightning Trigger is the only one I’d use and recommend (I get no kickback for this). On the other hand, if you already have a lightning sensor you’re happy with, there’s no reason to switch.
I won’t get into lots of specifics about how to set up the Lightning Trigger because it’s simple and covered fairly well in the included documentation. But you should know that of the things that sets the Lightning Trigger apart from many others is its ability to put your camera in the “shutter half pressed” mode, which greatly reduces shutter lag (see below). But that also means that connecting the Trigger will probably disable your LCD replay, so you won’t be able to review your captures without disconnecting—a simple but sometimes inconvenient task. You also probably won’t be able to adjust your exposure with the Lightning Trigger connected.
The Lightning Trigger documentation promises at least a 20 mile range, and after many years using mine at the Grand Canyon, I’ve seen nothing that causes me to question that. It also says you can expect the sensor to fire at lightning that’s not necessarily in front of you, or lightning you can’t see at all, which I will definitely confirm. For every click with lightning in my camera’s field of view, I get many clicks caused by lightning I didn’t see, or that were outside my camera’s field of view. But when visible lightning does fire somewhere in my composition, I estimate that the Lightning Trigger clicked the shutter at least 95 percent of the time (that is, even though I got lots of false positives, the Lightning Trigger missed very few bolts it should have detected). Of these successful clicks, I actually captured lightning in at least 2/3 of the frames.
The misses are a function of the timing between lightning and camera—sometimes the lightning is just too fast for the camera’s shutter lag. In general, the more violent the storm, the greater the likelihood of bolts of longer duration, and multiple strokes that are easier to capture. And my success rate has increased significantly beyond 2/3 since switching from a Canon 5DIII to Sony mirrorless (more on this in the Shutter Lag section).
The Lightning Trigger documentation recommends shutter speeds between 1/4 and 1/20 second—shutter speeds faster than 1/20 second risk completing the exposure before all of the secondary strokes fire; slower shutter speeds tend to wash out the lightning. To achieve daylight shutter speeds between 1/4 and 1/20 second, I use a polarizer, with my camera at ISO 50 and aperture at f/16 (and sometimes smaller). Of course exposure values will vary with the amount of light available, and you may not need such extreme settings when shooting into an extremely dark sky. The two stops of light lost to a polarizer helps a lot, and 4- or 6-stop neutral density filter is even better with fairly bright skies (but if you’re using a neutral density filter, try to avoid shutter speeds longer than 1/4 second).
Lightning is fast, really, really fast, so the faster your camera’s shutter responds after getting the command from the trigger device, the more success you’ll have. The delay between the click instruction (whether from your finger pressing the shutter button, a remote release, or a lightning sensor) and the shutter firing is called “shutter lag.”
The less shutter lag you have, the better your results will be. The two most important shutter lag factors are:
- Camera model: It’s surprising how much shutter lag can vary from manufacturer to manufacturer and model to model. In a perfect world, for lightning photography your camera’s shutter lag will be 60 milliseconds (.006 seconds) or faster (the lower the number the better), but 120 milliseconds (.012 seconds) or faster can give you some success. The top cameras from Sony, Nikon, and Canon are all fast enough, but the latest Sonys are the definite shutter lag winner (fastest), with Nikon a not too distant second, and Canon third. And shutter lag can vary with the manufacturer’s model: While my Sony a7RII is one of the fastest cameras out there, my a7R was unusably slow, so you need to check your model. Since I don’t check every camera released, it’s possible this ranking will change well before I update this article, so I recommend that you research shutter lag for your camera model. Unfortunately, shutter lag isn’t usually in the manufacturers specifications, so it’s hard to find. The best source I’ve found is the “Pre-focused” time in the Performance tab of the camera reviews at Imaging Resource.
- Camera settings: Basically, to minimize the “thinking” the camera needs to before firing, you want to be in manual everything mode—metering and focus. If your camera offers an electronic front curtain option (as my Sonys do), use it. If you must autofocus, go ahead and do it each time you recompose, then turn autofocus off as soon as you’re focused. Though the Lightning Trigger documentation suggests Aperture Priority metering, I use and recommend Manual metering mode to eliminate any camera-slowing metering (but Aperture Priority is fine if you have a strong preference). And, also despite what the Lightning Trigger documentation suggests, noise reduction is a post-capture function that might slightly delay continuous frames, but it won’t increase shutter lag.
In addition to a lightning sensor and fast camera, you’ll need:
- A solid tripod and head: Don’t even think about trying to photograph lightning hand-held
- Rain gear that keeps you dry from head-to-toe
- Umbrella (a.k.a., Wile E. Coyote Lightning Rod) to shield your camera and lightning sensor (many sensors, including the Lightning Trigger, aren’t waterproof) while you compose and wait in the rain. The umbrella is for when you’re photographing storm cells at a great distance, such as on the rim of the Grand Canyon and the lighting is across the canyon. Obviously, when the lightning gets within 10 miles, put the umbrella down and run for cover.)
- Lens hood to shield some of the raindrops that could mar the front element of your lenses
- Neutral density filter and/or polarizer to slow shutter speed into the ideal range (1/4 – 1/20 second)
- A garbage bag (my choice) or rainproof camera jacket (haven’t found one) to keep your camera and sensor dry during a downpour
- Extra lightning sensor batteries (better safe than sorry)
- Extra memory cards: When a storm is very close or active, your lightning sensor could detect 20 or 30 strikes per minute (even when little or no lightning is visible to the eye)
- Infrared remote to test your Lightning Trigger; I sometimes borrow the remote from my hotel room, but the Apple TV remote works great and is extremely compact (fits nicely into the Lightning Trigger pouch)
- A towel
Getting the shot
Lightning is most likely to strike in or near the gray curtains (clearly recognizable as distant rain) that hang beneath dark clouds. In addition to visible rain curtains, the darkest and tallest clouds are usually the most likely to fire lightning. Here are a few more points to consider:
- The wider your composition, the greater your odds of capturing lightning, but the smaller the lightning will appear in your image.
- Identify the most likely lightning cell and find the best composition that includes it. I tend to start with wider compositions to ensure success, then tighten my composition once I’m fairly confident I captured something.
- Note the height from which the lightning originates and be sure to include enough cloud to get all of the stroke. On the other hand, don’t include too much room above the lightning—the most frequent rookie mistake I see is too much sky/clouds in the frame. The second most frequent is lightning cut off at the top. Unless the storm is too close for safety, for any given cell, most lightning will originate from about the same height above the ground.
- The best lens is usually a midrange zoom such as a 24-70 or 24-105—if you find yourself reaching for the 16-35 (or wider), you’re too close.
- On the other hand, once you’re sure you’ve captured some good strikes, try putting on a 70-200. The narrow field of view can significantly reduce the number of frames with lightning, but the ones you get will be much larger in the frame and therefore more spectacular.
- Don’t forget to try some vertical compositions. I usually wait until after I know I’ve captured some in a horizontal frame because vertical narrows the horizontal field of view and lowers the odds of success a little.
- Lightning stands out better in a slightly underexposed image. My target shutter speed is usually 1/8 second (slow enough to include multiple pulses, but not so slow that I risk washing out the lightning). When the sky is relatively bright, dropping to 1/15 or even 1/20 second can make the lightning stand out better than 1/8 (but risks losing secondary strikes). Conversely, when the sky is extremely dark and the lightning is firing like crazy, extending to 1/4 second might increase your chances for multiple pulses.
- Just because you’re standing around waiting for things to happen, doesn’t mean there’s nothing to do. Keep your eyes glued to the sky and adjust your composition as the lightning shifts, or as new activity starts elsewhere. If you wait until you hear your shutter click or someone else exclaim before looking up, you won’t see the lightning. And monitor the light—your exposure can change by several stops as the storm moves, intensifies, or winds down.
- Try not to check your captures on your LCD until you’re done (or better yet, until you upload your images to your computer). With the Lightning Trigger (and some other sensors), viewing the LCD requires turning off the sensor, which risks missing a shot (I’m pretty sure lightning waits for me to turn off my sensor), and you’ll also find that many successful captures, especially wide compositions with a relatively bright sky, just aren’t that visible on an LCD viewed in daylight anyway.
Do as I say (not as I do)
Be aware that electrical storms can move quite quickly, so you need to monitor them closely. Sometimes this simply means adjusting your composition to account for shifting lightning; other times it means retreating to the car if the cell threatens your location. No shot is worth your life.
About this image
On the first evening of last year’s second Grand Canyon Monsoon photo workshop, Don Smith and I took the group to Point Imperial for a sunset shoot. Based on the forecast we had little hope for lightning, but one thing I’ve learned over the many years of photographing the monsoon here is that the forecast isn’t the final word. We got another reminder of this that evening.
The view from Point Imperial is both expansive and different from other Grand Canyon vistas, stretching east across the Painted Desert and north to the Vermillion Cliffs. As the group made their way down to the vista platform, in the corner of my I thought I a lighting strike far to the north. A second bolt confirmed my discovery and soon we had the entire group lined up with cameras pointed and triggers ready.
With everyone in business, I set up my tripod and attached my Lightning Trigger to my Sony a7RIII. Since this lightning was close to 30 miles away, maybe farther than any lightning I’ve tried to photograph, so I hauled out my Sony 100-400 GM lens and zoomed in as tight as I could. I didn’t have to wait long to confirm that my Lightning Trigger would catch strikes this distant—it didn’t hurt that these were massive bolts, many with multiple pulses and forks.
Everyone was thrilled, so thrilled that it didn’t immediately register that the storm was moving our direction. I started at 400mm, but by the time I captured this frame I was just a little more than 100mm. That’s still a pretty safe distance, but with night almost on us and another cell moving in from the east, we decided to take our winnings and go home.
One final note: If you check my exposure settings, you’ll see that my shutter speed here was .4 seconds, well outside the 1/20-1/4 second range I suggest. But if you look at the other settings, you’ll see that I’d opened up to f/7.1, and had cranked my ISO to 400, an indication that twilight was settling in. Successful lightning photograph is all about contrast, and the darker the sky, the better the bolt stands out, even in a longer exposure. Had we stayed past dark (and lived), we could have jettisoned the Lighting Triggers and used multi-second exposures.
Join Don Smith and me in our next Grand Canyon Monsoon Photo Workshop
Read my article in Outdoor Photographer magazine, Shooting the Monsoon
A Lightning Gallery
Click an image for a closer look and slide show. Refresh the window to reorder the display.
Posted on April 5, 2019
Last Monday seemed like the perfect day for a poppy shoot in the foothills. I had the afternoon wide open—with the California media buzzing about this year’s “superbloom,” plus a forecast promising ideal conditions (calm wind and thin clouds), I couldn’t help dreaming about my own images of poppy-saturated fields. What could possibly go wrong?
Getting on the road proved a little more problematic than anticipated, but by 2 p.m. I was on my way, encouraged forward by an occasional poppy beside the freeway. Adding to my optimism, the aforementioned clouds were just right: thick enough to diffuse the sunlight, but not so dark that they’d close the sun-loving poppies. I exited the freeway as soon as possible, opting to drive the 2-lane roads that follow the hills’ natural contours. While my preferred my route isn’t the most direct, it is the most scenic, winding me through oak-studded hills deeply greened by this year’s copious winter rain. Though this drive takes a little more than an hour, the time passes quickly with so much pastoral beauty filling my windshield.
I knew the poppies in Northern California were starting late due to our relatively late winter, but was fairly confident I’d allowed enough time for the golden hillsides to kick in. In a good spring, poppies dot the entire route, but by the time I was southbound on scenic Highway 49, I started realizing I hadn’t seen any poppies since leaving Sacramento. Soon I was pretty resigned to the fact that this year’s superbloom was limited Southern California, and wondered if I’d find any poppies at all. Then it started to rain.
As easy as it would have been easy to cut my losses and turn around, I simply changed my expectations. With fresh memories of a brief but rewarding raindrop experience in Yosemite, I realized I didn’t need to find entire hillsides covered with poppies, that even a single poppy could be nice. So, rather than zipping along Highway 49 at 50 MPH (-ish) looking for golden slopes, I started exploring some of the quieter tributary roads and quickly realized that there were a sprinkling of poppies out.
I ended up spending two hours photographing a small patch of poppies I found on a dead-end road near Jackson. It rained the entire time, but with rain gear in my car for just these situations, I stayed warm and dry. My camera? Not so much. I tried working with an umbrella, but after a few minutes realized I was one arm short and just decided to test the water resistance of my Sony a7RIII. I’m happy to say that it passed with flying colors, as did the Sony 100-400 GM.
In the two weeks since I shot those raindrops in Yosemite, I’ve been plotting how to get even closer. On the Yosemite shoot I added extension tubes to my 100-400; this afternoon I returned to the extension tubes, but added my 2X teleconverter (which, I might add, handled the rain perfectly as well). I thought I’d try a few lens/extension-tube/teleconverter configurations, but I was having so much fun that I ended up shooting this way the entire time.
On a rainy day, light is already limited. But adding a teleconverter and extension tubes compounds the light problem. Because f/stop is a ratio with focal length as the numerator and lens opening as the denominator, adding a teleconverter and extension increases the focal length, resulting in less light reaching the sensor. A 2x teleconverter cuts two stops of light, which means my 100-400 that’s normally wide upon f/5.6 at 400mm becomes f/11 at (the teleconverted) 800mm (400mm x 2). And adding extension tubes also extends the lens’s effective focal length, further reducing the light reaching the sensor. To compensate for all this missing light, I shot everything this afternoon at either ISO 1600 or ISO 3200.
One of the cool things about this kind of photography is how different the world looks through the viewfinder. I love putting my eye to the viewfinder, moving the lens around, and changing focus slowly to see what snaps into view. In this case I was looking for a poppy to isolate from its nearby surroundings, but that also has something nearby (usually another flower) that I could soften enough to complement without competing. Sometimes I had a general idea of a subject before looking through my camera, other times I’d just explore with my lens until something stopped me.
Because depth of field shrinks not only with focal length, but also with focus distance, every frame I clicked this afternoon had a paper-thin range of sharpness. With such a shallow depth of field, none of these images would have been possible without a tripod. With my composition set, I’d pick a focus point (usually, but not always, a prominent raindrop), focus in my viewfinder until I was “certain” it was sharp, then instantly debunk my that “certainty” by magnifying the image in my viewfinder. This little exercise quickly taught me that with such a small margin for error, the best I could reliably achieve without magnifying the view was almost sharp enough, making pre-click magnification an essential part of my focus workflow (instead of just a cursory focus-check).
Each time I do this kind of photography I learn something. In this case it was how far away I could be and still fill my frame with a poppy. All of the images I captured this afternoon were from four to six feet away.
I wrapped up when the sky darkened further and the rain started coming down pretty hard. I couldn’t believe I’d been out there two hours, and spent most of the drive strategizing new ideas for the next time.
Getting the Drop on Nature
Click an image for a closer look and to view slide show.