Deep learning-based multi-exposure image fusion (MEF) methods have demonstrated robust performance. However, these methods require considerable computational resources and energy, which greatly limits their practical deployment. To address this issue, we propose a lightweight, dual-stage MEF method, termed LDMEF. By effectively deploying on field-programmable gate array (FPGA), this method significantly enhances its range of applications and flexibility. Specifically, in the initial stage, LDMEF preprocesses the input sequences by leveraging the parallel processing capabilities of FPGA to compute a preliminary image through pixel-wise addition and averaging, ensuring both simple and rapid execution. Subsequently, in the second stage, our proposed method incorporates depthwise separable convolution with the preliminary image to facilitate a lightweight network that is both straightforward to deploy and simple in design. This network meticulously fine-tunes the preliminary image at the pixel level, achieving high-quality fusion results. Extensive evaluations on publicly available datasets confirm that LDMEF not only achieves remarkable results but also outperforms many GPU-based learning MEF methods.