struct circ_buf *xmit = &sport->port.state->xmit;
        unsigned long flags;
 
-       dma_unmap_sg(sport->port.dev, sgl, sport->dma_tx_nents, DMA_TO_DEVICE);
+       spin_lock_irqsave(&sport->port.lock, flags);
 
-       sport->dma_is_txing = 0;
+       dma_unmap_sg(sport->port.dev, sgl, sport->dma_tx_nents, DMA_TO_DEVICE);
 
        /* update the stat */
-       spin_lock_irqsave(&sport->port.lock, flags);
        xmit->tail = (xmit->tail + sport->tx_bytes) & (UART_XMIT_SIZE - 1);
        sport->port.icount.tx += sport->tx_bytes;
-       spin_unlock_irqrestore(&sport->port.lock, flags);
 
        dev_dbg(sport->port.dev, "we finish the TX DMA.\n");
 
+       sport->dma_is_txing = 0;
+
+       spin_unlock_irqrestore(&sport->port.lock, flags);
+
        uart_write_wakeup(&sport->port);
 
        if (waitqueue_active(&sport->dma_wait)) {
        struct dma_async_tx_descriptor *desc;
        struct dma_chan *chan = sport->dma_chan_tx;
        struct device *dev = sport->port.dev;
-       enum dma_status status;
        int ret;
 
-       status = dmaengine_tx_status(chan, (dma_cookie_t)0, NULL);
-       if (DMA_IN_PROGRESS == status)
+       if (sport->dma_is_txing)
                return;
 
        sport->tx_bytes = uart_circ_chars_pending(xmit);