A backup strategy for the images you edit in Lightroom Creative Cloud

My big 2025 photography project was to move all my pictures out of my hardware dependent local storage, and migrate them to Adobe’s Creative Cloud. I knew I would not convert my Lightroom 6 catalogs (theoretically possible, but too cumbersome), but the folders on the Network Attached Storage device (NAS) where the originals were stored had always been carefully organized. I thought I would not lose much by not converting the catalogs. I subscribed to Adobe’s Lightroom Creative Cloud (first through the Apple Store, later directly on the Adobe Store), and uploaded all the original images, folder by folder, to Lightroom. The process was described in detail is a series of blog entries dedicated to Lightroom.

Which means I’m now trusting Adobe for preserving 28 years of scanned negatives and digital images in their cloud. What can possibly go wrong?

A recent post by Jim Grey (about “the lost photos era”) and interactions I’ve had with cloud service providers in a professional context brought back to my attention that storing my images in a cloud was a good first step but not enough.

Rome – Fontana de Nettuno – Piazza Navona. Nikon D80 – Jan 2010

The “shared responsibility model”

All cloud service providers (CSPs) operate under a shared responsibility model. It’s the CSP’s job to ensure that their technical platform remains available and secure, and that the data entrusted to them can be recovered in case of a disaster in their data centers. As the client, it’s your responsibility to “govern your content”: manage the uploads, the regular cleanups, and configure how the data is accessed and shared.

The grey area is of course backup – CSPs generally commit to recovering your data at Day Minus One if something really bad happens to their infrastructure, but they won’t be obligated to do anything if you deleted a folder by mistake, or if you wanted to recover a group of files as they were at a specific point in time. CSPs generally consider that backups and restores are the responsibility of the client.

Although Adobe is a reliable company, I know I have to protect my images from a catastrophic error on their part, and from a major mistake (fat finger?) on mine.

Rome, Jan 2010 – Nikon D80

A reminder – the differences between Adobe Lightroom and Adobe Lightroom Classic

Adobe Lightroom Classic is the current iteration of Adobe’s original image edit and management software, launched in 2007 as Lightroom 1.0.

It’s a “fat client” application designed to work on Windows or MacOS workstations (desktop or laptop), which stores your images locally (on the hard drive of your workstation or on some form of higher capacity local storage, DAS or NAS). Lightroom maintains at least one local catalog of your images, which contains all the ratings, flags, titles, captions you have entered, as well as a log of all the edits and setting changes (crops, exposure, color balance, sharpening,…) performed on the images.

The system is totally self contained – but as everything (catalog, images and edits) is kept locally, it’s your responsibility to manage the storage, the backup and the disaster recovery of your images.

Under the same Lightroom brand, Adobe is selling a totally different range of cloud based products simply named Lightroom or Lightroom CC, whose lightweight clients run on a smartphone (iOS or Android), a tablet (iPadOS or Android) and on a desktop or laptop (Windows or MacOS). All those products share the same on-line library (hosted on Adobe’s Creative Cloud).

Contrarily to Lightroom Classic, the Creative Cloud versions of Lightroom (smartphone, tablet, PC or Mac) don’t keep any image or catalog on your device – just a cache to reduce the response time. The whole system works very well: I can upload images from my camera through a smartphone while traveling, perform light edits on a tablet at the hotel the same day, and spend more time perfecting the images on a laptop when I’m back home – it’s seamless. As long as I keep paying for the subscription, of course. And bar a catastrophic event in Creative Cloud.

Rome – Fontana de Nettuno – Piazza Navona. Nikon D80 – Jan 2010

Backup workflows don’t live forever

Even if the image formats themselves (jPEG and DNG) have been remarkably stable over the last 20 years, the hardware, the software and the cloud services offerings have not stopped evolving – and what used to work reliably ten years ago does not work any more. Which means that every now and then, we need to take a hard look at our workflow and re-engineer it.

When I put it in place in 2018, my image preservation workflow made sense – I was using Adobe Lightroom 6 running on a Mac to edit my photos and manage my libraries. Lightroom 6 was keeping the catalog on the local hard drive of my Macbook and was pointing to a volume on the Netgear NAS to store the images themselves. I was also running a backup application named Arq on the Macbook, and using it to keep a backup of the NAS in Amazon’s long term storage, AWS Glacier.

Along the years, this finely tuned workflow crumbled.

First, the OS of my old MacBook stopped being supported, and I saw its capabilities decline progressively as it could not access the services that Apple (and others) kept on making more secure with more refined security protocols and longer encryption keys.

To make the matter worse, Netgear decided to get out of the network storage business – my RN214 NAS still works, but is not supported and (of course) its OS and its built-in backup apps are not updated anymore.

Last but not least, AWS has now sunset Glacier as I was using it – it’s not a stand alone product anymore, just a storage class in the S3 product portfolio, using different APIs.

Rome, Jan 2010 – Nikon D80

My storage and backup strategy was crumbling and I had to act. That’s why I migrated the libraries themselves to Adobe’s Creative Cloud last year, and why I’m now implementing a new backup and restore workflow now.

My new workflow – saving the “digital negatives

As often nowadays, I called ChatGPT for help. The workflow it recommended, and that I implemented, is still based on Adobe Creative Cloud being my primary image store, the “source of truth”. Lightroom (the PC/Mac edition of Lightroom) on my MacBook will act as a sort of gateway to the NAS, and the NAS volume will store my local replica of the originals stored in Creative Cloud.

It’s important to remember that for Lightroom, a local storage volume is nothing more than the place where it stores a local cache. What is being replicated to the local volume is the source image – the original JPEG or raw files exactly as they were originally uploaded from the camera – before any transformation, optimization or edit was performed. The images are grouped on the SAN by date (one folder per year, one subfolder per day) and the album structure you defined in Lightroom is not respected. Again, it’s a cache that we use as a way to backup our source images, not a backup of the final images after Lightroom has processed them.

The local cache on the SAN shows the original files grouped by date of capture – the Lightroom Album structure and the edits are not preserved, only the original image itself (compare with the structure of April 2016 in Lightroom, as shown below).
Lightroom CC – the folder/album structure (here, April 2016). In Lightroom the images are grouped in user defined folders and albums.

How to setup Adobe Lightroom

Once the Mac is logged in the Network Attached Storage volume, simply click on the “Adobe Lightroom” option at the top left of the screen, select “Cache”, and under Performance, check the “Store a copy of all originals option”, and point to the folder of the NAS where the original images will be dropped.

The sync process is managed automatically by Lightroom. Every time you add new pictures to Lightroom, it will start replicating them to the SAN.

If you’re working with Lightroom away from your home network, no problem. Adobe will consider that the cache is not available, and will download the images from the cloud.

In Lightroom CC – check the “Store a copy of all originals” option and point to the NAS as a the local storage

Creating an off-site backup of the Network Attached Storage volume

The primary storage location of my images is Adobe Creative Cloud. I keep a replica of the originals on a network attached storage device (NAS) at home. It’s a pretty solid data protection system, but it’s only keeping one replica of Creative Cloud’s originals – and a replica is not a backup (because it only keeps the most recent version of a file). It is not very complicated or expensive to make it even more robust, and create an off site backup of the original images.

Duplicati – the backup job (it took it 12 hours to backup 110 Gbytes of pictures – not bad at all).

That’s what I used to do with Amazon Glacier – and having an off site backup of my photo library was a saving grace when my first Netgear NAS device gave up the ghost. Restoring the images from Glacier took a week, but it’s better than losing everything.

The target Google Drive after the backup – the data is grouped in blocks of 50 Mbytes.

This time, I tried different options (Arq, Backblaze) which for various reasons (performance, cost, no support of network attached devices) did not work for me. My current setup is based on an open source software named Duplicati, which is pushing the Lightroom replica on the NAS to a Google Drive. It works, backups are reasonably fast (around 2.5 Mbytes/sec), and it’s flexible enough: I can recover a specific image in a few minutes if I need to.

Validating that the setup works

Backup and restore workflows are fragile, and they can fail for all sorts of reasons (expired passwords or keys, OS or software upgrade, hardware or network related issues, human error). And it’s not because the backup is successful that the restore will be.

Restoring the data – selecting the document to restore is easy and the restore takes no more than a few minutes.

I had to validate that, with the Mac and Lightroom CC up and running, and the NAS volume mounted, that:

  • any new image added to the Lightroom CC library was replicated to the SAN, in its original state,
  • the backup software would catch the new image and back it up to the Google drive,
  • and that I could restore any image or any group of images as needed.

The tests were successful.

Saving the final images

You may also want to preserve a copy of the final state of your images, after Lightroom has applied all of its edits.

The challenge of course is that in Lightroom, the images don’t really have a final state. Adobe keeps your original photo and a sort of log of the transformations you performed, and dynamically creates a file containing the image you want after you have requested an export. You pick the quality, the dimensions, and the file format (small JPEG, large JPEG, PNG, TIFF, DNG, …) with or without sharpening – depending on what you intend to do with the image (email attachment, social media, photo gallery, photo album, print, …). And the image you need is created on the fly.

Lightroom – so many ways to export a photo

I understand that a professional photographer delivering images to many clients may want to keep a trace of what was delivered, and have an archival system specifically tuned to preserve them. (And pros may prefer working Adobe Lightroom Classic, anyway).

I’m not in this situation and I’ve never really given much thought about it. I simply export the images I need to the same shared folder in Apple’s iCloud, that all my Apple devices (iPhone, iPad, MacBook) can access.

Final words

In the days of film, it was not easy (or cheap) for amateur photographers to create duplicates of their color slides or their negatives, and store them in a second location as a backup. Photographers were at the mercy of fire, floods and burglaries, and could lose the images of a lifetime in a few minutes.

Digital images can be easily duplicated, and the duplicates stored in totally different locations, on totally different media. The setup described here is very easy to implement: a NAS is not even needed (the local SSD of a PC or a Mac would work as well), and many of our subscriptions (Microsoft Office Family for instance) already include 1 TB of storage and could be used as backup target.


More about Lightroom and fifty five camera reviews:


All the images of this series were shot in Rome, in April 2009 and in January 2010, with a Nikon D80. They were saved as originals on multiple generations of storage, recovered from a catastrophic NAS failure, and imported in Adobe’s Creative Cloud last year. I just adjusted a few sliders before exporting them to WordPress.

Rome, Jan 2010 – Nikon D80
Rome, Jan 2010 – Nikon D80

Old iMacs never die, they just go slower

I recently moved to a new home – half the size of the one I was leaving. In a big home, you have lots of storage, a basement, an attic, a large garage, rooms you never use, and you tend to keep stuff. Lots of old stuff. Like old computers.

imacG4
The iMac G4 fully deployed with keyboard, mouse and loudspeakers. Source: unknown (probably a scan of an Apple catalog)

But when you move, you can’t leave anything behind. You check all the hidden corners of your home, and you find your old iMac. Second generation, officially known as the iMac G4 700 MHz. And often nicknamed Sunflower, because of its very original shape.

I had bought it in 2002. My first Mac. My first Apple widget.

I decided to give it a test run. In the spirit of “a new life for old gear”. But if 40 year cameras can still produce nice pictures (provided you can find film for them, of course), computers half their age are hardly usable.

The machine did age very well. It had been solidly built, with good quality components. It boots, and makes all the right sounds and moves. You log-in, and find a desktop which is not really different from what you see on a modern Mac (this machine is running OS X 10.5 Leopard – the last OS X version to run over PowerPC processors). Your spirits are high. You’ve been reunited with an old friend.

On paper, the specs of an old iMac – a 700 MHz 32 bit processor, 512 MB RAM, WiFi, Firewire – are not totally ridiculous if you compare them with a recent entry level machine. A modern Macbook has a 64 bit dual core processor, generally running at 1.3 GHz, with the ability to burst at 3 GHz. It does not look like a huge difference. Admittedly, a modern laptop also has much more RAM (8 GB) and more storage (256 GB SSD as opposed to a 40 GB spinning drive). But in the real life, it results in a huge gap in speed.

IMG_2703
Technical specs of the iMac G4 700 MHz (from the package)

The iMac is incredibly slow at performing tasks as simple as ripping a music CD. And the worst is not the speed, it’s the inability to do anything really useful with the machine because the hardware and OS X Leopard can’t cope with modern security protocols:

  • you can’t connect to a wireless LAN, because the iMac and OS X Leopard support at best WPA (original version) and not WPA2 which is the standard today.
  • you can’t browse any modern Internet site, because the browser does not support the recent encryption layers of https, and every site worth its salt defaults to https now.
  • you can’t access any of the on-line services requiring a fat client (the Apple Music store from iTunes, for instance) because nobody’s accepting connections from such an old thing (probably another TLS vs SSL issue)
  • Software from 2009 still works (things like Photoshop Elements 2 or Microsoft Word   2008) but obviously recent versions of popular software are written for Intel Macs, and don’t run over Power PC processors (not even software as basic as a browser).

IMG_2704
iMac G4 – trying to browse an https page – fairly typical error message

Basically, it could work if you downgraded your home WiFi network to WPA (or even worse, WEP), and only browsed sites whose servers have not been updated since the end of support of OS X Leopard (in 2009, I believe). That’s a scary perspective.

So, what will happen with this Mac?  I will donate it. To people who still have room in their large home, where it will join a burgeoning computer museum. Until they decide to downsize of course.


An anecdote – the iMac was the first Apple computer I bought, but not the first I used. In the early eighties, I even had the privilege of being trained on Apple SOS (nicknamed Applesauce, of course), the operating system of the Apple III. It makes me a true geek.

For the fans of all things Apple, and among all the bios of Steve Jobs, I recommend the excellent “Becoming Steve Jobs“, by Brent Schlender and Rick Tetzeli.


IMG_2615
“les petits chapeaux” – Apple iPhone 7. Even a recent iPhone outperforms an iMac G4 – and it also takes pictures.

Lobbying to make the use of film mandatory?


White Elephant - Hollywood Boulevard - Los Angeles
The symbol of an industry? Two white elephants sit at the top of columns on Hollywood Boulevard.


You may have read about this last week – the immensely popular RIAA and the always forward looking NAB, representing respectively our beloved recording industry and the owners of radio stations, are teaming to lobby the US Congress in order to make FM radio receivers mandatory on every cell phone or smartphone sold in this country.


Cell phones have been available for more than 15 years now, and the few manufacturers who tried to sell cell phones with built in FM radio did not see any explosion of their sales volume. Ten years ago, the buyers of cell phones didn’t see the need of an integrated FM receiver, and now that Internet radios and Pandora are available, there are even less reasons to place a good old FM tuner in a brand new 3G smartphone.


By the way, what’s the justification for putting an FM receiver in a cell phone? Public safety. According to the NAB, cell phone users would be able to listen to emergency messages on their favorite FM radio station.


Too bad the photographic film industry did not have the same imagination. Or they could have imposed 35mm Film Cameras in candy bar cell phones, and Instant Film cameras in smartphones. Cell phone carriers would have obliged the film industry by proposing two year agreements including two new cartridges of film per month. And the justification would have been national security, of course.


Imagine the business opportunities. Apple negotiating a 5 year exclusivity with Polaroid, Verizon smartphones printing two copies of the same picture for the price of one. And another carrier imposing hefty fees for the consumers who did not burn their 48 pictures during the last 30 days.


We really missed something.