diff options
Diffstat (limited to 'drivers/gpu/drm/radeon/r600.c')
| -rw-r--r-- | drivers/gpu/drm/radeon/r600.c | 293 | 
1 files changed, 232 insertions, 61 deletions
diff --git a/drivers/gpu/drm/radeon/r600.c b/drivers/gpu/drm/radeon/r600.c index ea4d3734e6d..3c69f58e46e 100644 --- a/drivers/gpu/drm/radeon/r600.c +++ b/drivers/gpu/drm/radeon/r600.c @@ -105,6 +105,7 @@ void r600_fini(struct radeon_device *rdev);  void r600_irq_disable(struct radeon_device *rdev);  static void r600_pcie_gen2_enable(struct radeon_device *rdev);  extern int evergreen_rlc_resume(struct radeon_device *rdev); +extern void rv770_set_clk_bypass_mode(struct radeon_device *rdev);  /**   * r600_get_xclk - get the xclk @@ -119,6 +120,64 @@ u32 r600_get_xclk(struct radeon_device *rdev)  	return rdev->clock.spll.reference_freq;  } +int r600_set_uvd_clocks(struct radeon_device *rdev, u32 vclk, u32 dclk) +{ +	return 0; +} + +void dce3_program_fmt(struct drm_encoder *encoder) +{ +	struct drm_device *dev = encoder->dev; +	struct radeon_device *rdev = dev->dev_private; +	struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); +	struct radeon_crtc *radeon_crtc = to_radeon_crtc(encoder->crtc); +	struct drm_connector *connector = radeon_get_connector_for_encoder(encoder); +	int bpc = 0; +	u32 tmp = 0; +	enum radeon_connector_dither dither = RADEON_FMT_DITHER_DISABLE; + +	if (connector) { +		struct radeon_connector *radeon_connector = to_radeon_connector(connector); +		bpc = radeon_get_monitor_bpc(connector); +		dither = radeon_connector->dither; +	} + +	/* LVDS FMT is set up by atom */ +	if (radeon_encoder->devices & ATOM_DEVICE_LCD_SUPPORT) +		return; + +	/* not needed for analog */ +	if ((radeon_encoder->encoder_id == ENCODER_OBJECT_ID_INTERNAL_KLDSCP_DAC1) || +	    (radeon_encoder->encoder_id == ENCODER_OBJECT_ID_INTERNAL_KLDSCP_DAC2)) +		return; + +	if (bpc == 0) +		return; + +	switch (bpc) { +	case 6: +		if (dither == RADEON_FMT_DITHER_ENABLE) +			/* XXX sort out optimal dither settings */ +			tmp |= FMT_SPATIAL_DITHER_EN; +		else +			tmp |= FMT_TRUNCATE_EN; +		break; +	case 8: +		if (dither == RADEON_FMT_DITHER_ENABLE) +			/* XXX sort out optimal dither settings */ +			tmp |= (FMT_SPATIAL_DITHER_EN | FMT_SPATIAL_DITHER_DEPTH); +		else +			tmp |= (FMT_TRUNCATE_EN | FMT_TRUNCATE_DEPTH); +		break; +	case 10: +	default: +		/* not needed */ +		break; +	} + +	WREG32(FMT_BIT_DEPTH_CONTROL + radeon_crtc->crtc_offset, tmp); +} +  /* get temperature in millidegrees */  int rv6xx_get_temp(struct radeon_device *rdev)  { @@ -1045,20 +1104,27 @@ int r600_mc_wait_for_idle(struct radeon_device *rdev)  uint32_t rs780_mc_rreg(struct radeon_device *rdev, uint32_t reg)  { +	unsigned long flags;  	uint32_t r; +	spin_lock_irqsave(&rdev->mc_idx_lock, flags);  	WREG32(R_0028F8_MC_INDEX, S_0028F8_MC_IND_ADDR(reg));  	r = RREG32(R_0028FC_MC_DATA);  	WREG32(R_0028F8_MC_INDEX, ~C_0028F8_MC_IND_ADDR); +	spin_unlock_irqrestore(&rdev->mc_idx_lock, flags);  	return r;  }  void rs780_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v)  { +	unsigned long flags; + +	spin_lock_irqsave(&rdev->mc_idx_lock, flags);  	WREG32(R_0028F8_MC_INDEX, S_0028F8_MC_IND_ADDR(reg) |  		S_0028F8_MC_IND_WR_EN(1));  	WREG32(R_0028FC_MC_DATA, v);  	WREG32(R_0028F8_MC_INDEX, 0x7F); +	spin_unlock_irqrestore(&rdev->mc_idx_lock, flags);  }  static void r600_mc_program(struct radeon_device *rdev) @@ -1579,6 +1645,67 @@ static void r600_gpu_soft_reset(struct radeon_device *rdev, u32 reset_mask)  	r600_print_gpu_status_regs(rdev);  } +static void r600_gpu_pci_config_reset(struct radeon_device *rdev) +{ +	struct rv515_mc_save save; +	u32 tmp, i; + +	dev_info(rdev->dev, "GPU pci config reset\n"); + +	/* disable dpm? */ + +	/* Disable CP parsing/prefetching */ +	if (rdev->family >= CHIP_RV770) +		WREG32(R_0086D8_CP_ME_CNTL, S_0086D8_CP_ME_HALT(1) | S_0086D8_CP_PFP_HALT(1)); +	else +		WREG32(R_0086D8_CP_ME_CNTL, S_0086D8_CP_ME_HALT(1)); + +	/* disable the RLC */ +	WREG32(RLC_CNTL, 0); + +	/* Disable DMA */ +	tmp = RREG32(DMA_RB_CNTL); +	tmp &= ~DMA_RB_ENABLE; +	WREG32(DMA_RB_CNTL, tmp); + +	mdelay(50); + +	/* set mclk/sclk to bypass */ +	if (rdev->family >= CHIP_RV770) +		rv770_set_clk_bypass_mode(rdev); +	/* disable BM */ +	pci_clear_master(rdev->pdev); +	/* disable mem access */ +	rv515_mc_stop(rdev, &save); +	if (r600_mc_wait_for_idle(rdev)) { +		dev_warn(rdev->dev, "Wait for MC idle timedout !\n"); +	} + +	/* BIF reset workaround.  Not sure if this is needed on 6xx */ +	tmp = RREG32(BUS_CNTL); +	tmp |= VGA_COHE_SPEC_TIMER_DIS; +	WREG32(BUS_CNTL, tmp); + +	tmp = RREG32(BIF_SCRATCH0); + +	/* reset */ +	radeon_pci_config_reset(rdev); +	mdelay(1); + +	/* BIF reset workaround.  Not sure if this is needed on 6xx */ +	tmp = SOFT_RESET_BIF; +	WREG32(SRBM_SOFT_RESET, tmp); +	mdelay(1); +	WREG32(SRBM_SOFT_RESET, 0); + +	/* wait for asic to come out of reset */ +	for (i = 0; i < rdev->usec_timeout; i++) { +		if (RREG32(CONFIG_MEMSIZE) != 0xffffffff) +			break; +		udelay(1); +	} +} +  int r600_asic_reset(struct radeon_device *rdev)  {  	u32 reset_mask; @@ -1588,10 +1715,17 @@ int r600_asic_reset(struct radeon_device *rdev)  	if (reset_mask)  		r600_set_bios_scratch_engine_hung(rdev, true); +	/* try soft reset */  	r600_gpu_soft_reset(rdev, reset_mask);  	reset_mask = r600_gpu_check_soft_reset(rdev); +	/* try pci config reset */ +	if (reset_mask && radeon_hard_reset) +		r600_gpu_pci_config_reset(rdev); + +	reset_mask = r600_gpu_check_soft_reset(rdev); +  	if (!reset_mask)  		r600_set_bios_scratch_engine_hung(rdev, false); @@ -1614,11 +1748,9 @@ bool r600_gfx_is_lockup(struct radeon_device *rdev, struct radeon_ring *ring)  	if (!(reset_mask & (RADEON_RESET_GFX |  			    RADEON_RESET_COMPUTE |  			    RADEON_RESET_CP))) { -		radeon_ring_lockup_update(ring); +		radeon_ring_lockup_update(rdev, ring);  		return false;  	} -	/* force CP activities */ -	radeon_ring_force_activity(rdev, ring);  	return radeon_ring_test_lockup(rdev, ring);  } @@ -1826,6 +1958,9 @@ static void r600_gpu_init(struct radeon_device *rdev)  	if (tmp < rdev->config.r600.max_simds) {  		rdev->config.r600.max_simds = tmp;  	} +	tmp = rdev->config.r600.max_simds - +		r600_count_pipe_bits((cc_gc_shader_pipe_config >> 16) & R6XX_MAX_SIMDS_MASK); +	rdev->config.r600.active_simds = tmp;  	disabled_rb_mask = (RREG32(CC_RB_BACKEND_DISABLE) >> 16) & R6XX_MAX_BACKENDS_MASK;  	tmp = (tiling_config & PIPE_TILING__MASK) >> PIPE_TILING__SHIFT; @@ -2092,20 +2227,27 @@ static void r600_gpu_init(struct radeon_device *rdev)   */  u32 r600_pciep_rreg(struct radeon_device *rdev, u32 reg)  { +	unsigned long flags;  	u32 r; +	spin_lock_irqsave(&rdev->pciep_idx_lock, flags);  	WREG32(PCIE_PORT_INDEX, ((reg) & 0xff));  	(void)RREG32(PCIE_PORT_INDEX);  	r = RREG32(PCIE_PORT_DATA); +	spin_unlock_irqrestore(&rdev->pciep_idx_lock, flags);  	return r;  }  void r600_pciep_wreg(struct radeon_device *rdev, u32 reg, u32 v)  { +	unsigned long flags; + +	spin_lock_irqsave(&rdev->pciep_idx_lock, flags);  	WREG32(PCIE_PORT_INDEX, ((reg) & 0xff));  	(void)RREG32(PCIE_PORT_INDEX);  	WREG32(PCIE_PORT_DATA, (v));  	(void)RREG32(PCIE_PORT_DATA); +	spin_unlock_irqrestore(&rdev->pciep_idx_lock, flags);  }  /* @@ -2113,7 +2255,8 @@ void r600_pciep_wreg(struct radeon_device *rdev, u32 reg, u32 v)   */  void r600_cp_stop(struct radeon_device *rdev)  { -	radeon_ttm_set_active_vram_size(rdev, rdev->mc.visible_vram_size); +	if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX) +		radeon_ttm_set_active_vram_size(rdev, rdev->mc.visible_vram_size);  	WREG32(R_0086D8_CP_ME_CNTL, S_0086D8_CP_ME_HALT(1));  	WREG32(SCRATCH_UMSK, 0);  	rdev->ring[RADEON_RING_TYPE_GFX_INDEX].ready = false; @@ -2283,6 +2426,7 @@ int r600_init_microcode(struct radeon_device *rdev)  			       fw_name);  			release_firmware(rdev->smc_fw);  			rdev->smc_fw = NULL; +			err = 0;  		} else if (rdev->smc_fw->size != smc_req_size) {  			printk(KERN_ERR  			       "smc: Bogus length %zu in firmware \"%s\"\n", @@ -2309,6 +2453,36 @@ out:  	return err;  } +u32 r600_gfx_get_rptr(struct radeon_device *rdev, +		      struct radeon_ring *ring) +{ +	u32 rptr; + +	if (rdev->wb.enabled) +		rptr = rdev->wb.wb[ring->rptr_offs/4]; +	else +		rptr = RREG32(R600_CP_RB_RPTR); + +	return rptr; +} + +u32 r600_gfx_get_wptr(struct radeon_device *rdev, +		      struct radeon_ring *ring) +{ +	u32 wptr; + +	wptr = RREG32(R600_CP_RB_WPTR); + +	return wptr; +} + +void r600_gfx_set_wptr(struct radeon_device *rdev, +		       struct radeon_ring *ring) +{ +	WREG32(R600_CP_RB_WPTR, ring->wptr); +	(void)RREG32(R600_CP_RB_WPTR); +} +  static int r600_cp_load_microcode(struct radeon_device *rdev)  {  	const __be32 *fw_data; @@ -2431,8 +2605,6 @@ int r600_cp_resume(struct radeon_device *rdev)  	WREG32(CP_RB_BASE, ring->gpu_addr >> 8);  	WREG32(CP_DEBUG, (1 << 27) | (1 << 28)); -	ring->rptr = RREG32(CP_RB_RPTR); -  	r600_cp_start(rdev);  	ring->ready = true;  	r = radeon_ring_test(rdev, RADEON_RING_TYPE_GFX_INDEX, ring); @@ -2440,6 +2612,10 @@ int r600_cp_resume(struct radeon_device *rdev)  		ring->ready = false;  		return r;  	} + +	if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX) +		radeon_ttm_set_active_vram_size(rdev, rdev->mc.real_vram_size); +  	return 0;  } @@ -2534,30 +2710,31 @@ void r600_fence_ring_emit(struct radeon_device *rdev,  			  struct radeon_fence *fence)  {  	struct radeon_ring *ring = &rdev->ring[fence->ring]; +	u32 cp_coher_cntl = PACKET3_TC_ACTION_ENA | PACKET3_VC_ACTION_ENA | +		PACKET3_SH_ACTION_ENA; + +	if (rdev->family >= CHIP_RV770) +		cp_coher_cntl |= PACKET3_FULL_CACHE_ENA;  	if (rdev->wb.use_event) {  		u64 addr = rdev->fence_drv[fence->ring].gpu_addr;  		/* flush read cache over gart */  		radeon_ring_write(ring, PACKET3(PACKET3_SURFACE_SYNC, 3)); -		radeon_ring_write(ring, PACKET3_TC_ACTION_ENA | -					PACKET3_VC_ACTION_ENA | -					PACKET3_SH_ACTION_ENA); +		radeon_ring_write(ring, cp_coher_cntl);  		radeon_ring_write(ring, 0xFFFFFFFF);  		radeon_ring_write(ring, 0);  		radeon_ring_write(ring, 10); /* poll interval */  		/* EVENT_WRITE_EOP - flush caches, send int */  		radeon_ring_write(ring, PACKET3(PACKET3_EVENT_WRITE_EOP, 4));  		radeon_ring_write(ring, EVENT_TYPE(CACHE_FLUSH_AND_INV_EVENT_TS) | EVENT_INDEX(5)); -		radeon_ring_write(ring, addr & 0xffffffff); +		radeon_ring_write(ring, lower_32_bits(addr));  		radeon_ring_write(ring, (upper_32_bits(addr) & 0xff) | DATA_SEL(1) | INT_SEL(2));  		radeon_ring_write(ring, fence->seq);  		radeon_ring_write(ring, 0);  	} else {  		/* flush read cache over gart */  		radeon_ring_write(ring, PACKET3(PACKET3_SURFACE_SYNC, 3)); -		radeon_ring_write(ring, PACKET3_TC_ACTION_ENA | -					PACKET3_VC_ACTION_ENA | -					PACKET3_SH_ACTION_ENA); +		radeon_ring_write(ring, cp_coher_cntl);  		radeon_ring_write(ring, 0xFFFFFFFF);  		radeon_ring_write(ring, 0);  		radeon_ring_write(ring, 10); /* poll interval */ @@ -2577,7 +2754,7 @@ void r600_fence_ring_emit(struct radeon_device *rdev,  	}  } -void r600_semaphore_ring_emit(struct radeon_device *rdev, +bool r600_semaphore_ring_emit(struct radeon_device *rdev,  			      struct radeon_ring *ring,  			      struct radeon_semaphore *semaphore,  			      bool emit_wait) @@ -2589,8 +2766,10 @@ void r600_semaphore_ring_emit(struct radeon_device *rdev,  		sel |= PACKET3_SEM_WAIT_ON_SIGNAL;  	radeon_ring_write(ring, PACKET3(PACKET3_MEM_SEMAPHORE, 1)); -	radeon_ring_write(ring, addr & 0xffffffff); +	radeon_ring_write(ring, lower_32_bits(addr));  	radeon_ring_write(ring, (upper_32_bits(addr) & 0xff) | sel); + +	return true;  }  /** @@ -2633,13 +2812,8 @@ int r600_copy_cpdma(struct radeon_device *rdev,  		return r;  	} -	if (radeon_fence_need_sync(*fence, ring->idx)) { -		radeon_semaphore_sync_rings(rdev, sem, (*fence)->ring, -					    ring->idx); -		radeon_fence_note_sync(*fence, ring->idx); -	} else { -		radeon_semaphore_free(rdev, &sem, NULL); -	} +	radeon_semaphore_sync_to(sem, *fence); +	radeon_semaphore_sync_rings(rdev, sem, ring->idx);  	radeon_ring_write(ring, PACKET3(PACKET3_SET_CONFIG_REG, 1));  	radeon_ring_write(ring, (WAIT_UNTIL - PACKET3_SET_CONFIG_REG_OFFSET) >> 2); @@ -2653,9 +2827,9 @@ int r600_copy_cpdma(struct radeon_device *rdev,  		if (size_in_bytes == 0)  			tmp |= PACKET3_CP_DMA_CP_SYNC;  		radeon_ring_write(ring, PACKET3(PACKET3_CP_DMA, 4)); -		radeon_ring_write(ring, src_offset & 0xffffffff); +		radeon_ring_write(ring, lower_32_bits(src_offset));  		radeon_ring_write(ring, tmp); -		radeon_ring_write(ring, dst_offset & 0xffffffff); +		radeon_ring_write(ring, lower_32_bits(dst_offset));  		radeon_ring_write(ring, upper_32_bits(dst_offset) & 0xff);  		radeon_ring_write(ring, cur_size_in_bytes);  		src_offset += cur_size_in_bytes; @@ -2668,6 +2842,7 @@ int r600_copy_cpdma(struct radeon_device *rdev,  	r = radeon_fence_emit(rdev, fence, ring->idx);  	if (r) {  		radeon_ring_unlock_undo(rdev, ring); +		radeon_semaphore_free(rdev, &sem, NULL);  		return r;  	} @@ -2705,14 +2880,6 @@ static int r600_startup(struct radeon_device *rdev)  	r600_mc_program(rdev); -	if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw) { -		r = r600_init_microcode(rdev); -		if (r) { -			DRM_ERROR("Failed to load firmware!\n"); -			return r; -		} -	} -  	if (rdev->flags & RADEON_IS_AGP) {  		r600_agp_enable(rdev);  	} else { @@ -2733,12 +2900,6 @@ static int r600_startup(struct radeon_device *rdev)  		return r;  	} -	r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_DMA_INDEX); -	if (r) { -		dev_err(rdev->dev, "failed initializing DMA fences (%d).\n", r); -		return r; -	} -  	/* Enable IRQ */  	if (!rdev->irq.installed) {  		r = radeon_irq_kms_init(rdev); @@ -2756,18 +2917,10 @@ static int r600_startup(struct radeon_device *rdev)  	ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];  	r = radeon_ring_init(rdev, ring, ring->ring_size, RADEON_WB_CP_RPTR_OFFSET, -			     R600_CP_RB_RPTR, R600_CP_RB_WPTR,  			     RADEON_CP_PACKET2);  	if (r)  		return r; -	ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; -	r = radeon_ring_init(rdev, ring, ring->ring_size, R600_WB_DMA_RPTR_OFFSET, -			     DMA_RB_RPTR, DMA_RB_WPTR, -			     DMA_PACKET(DMA_PACKET_NOP, 0, 0, 0)); -	if (r) -		return r; -  	r = r600_cp_load_microcode(rdev);  	if (r)  		return r; @@ -2775,10 +2928,6 @@ static int r600_startup(struct radeon_device *rdev)  	if (r)  		return r; -	r = r600_dma_resume(rdev); -	if (r) -		return r; -  	r = radeon_ib_pool_init(rdev);  	if (r) {  		dev_err(rdev->dev, "IB initialization failed (%d).\n", r); @@ -2819,6 +2968,9 @@ int r600_resume(struct radeon_device *rdev)  	/* post card */  	atom_asic_init(rdev->mode_info.atom_context); +	if (rdev->pm.pm_method == PM_METHOD_DPM) +		radeon_pm_resume(rdev); +  	rdev->accel_working = true;  	r = r600_startup(rdev);  	if (r) { @@ -2832,9 +2984,9 @@ int r600_resume(struct radeon_device *rdev)  int r600_suspend(struct radeon_device *rdev)  { +	radeon_pm_suspend(rdev);  	r600_audio_fini(rdev);  	r600_cp_stop(rdev); -	r600_dma_stop(rdev);  	r600_irq_suspend(rdev);  	radeon_wb_disable(rdev);  	r600_pcie_gart_disable(rdev); @@ -2900,12 +3052,20 @@ int r600_init(struct radeon_device *rdev)  	if (r)  		return r; +	if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw) { +		r = r600_init_microcode(rdev); +		if (r) { +			DRM_ERROR("Failed to load firmware!\n"); +			return r; +		} +	} + +	/* Initialize power management */ +	radeon_pm_init(rdev); +  	rdev->ring[RADEON_RING_TYPE_GFX_INDEX].ring_obj = NULL;  	r600_ring_init(rdev, &rdev->ring[RADEON_RING_TYPE_GFX_INDEX], 1024 * 1024); -	rdev->ring[R600_RING_TYPE_DMA_INDEX].ring_obj = NULL; -	r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_DMA_INDEX], 64 * 1024); -  	rdev->ih.ring_obj = NULL;  	r600_ih_ring_init(rdev, 64 * 1024); @@ -2918,7 +3078,6 @@ int r600_init(struct radeon_device *rdev)  	if (r) {  		dev_err(rdev->dev, "disabling GPU acceleration\n");  		r600_cp_fini(rdev); -		r600_dma_fini(rdev);  		r600_irq_fini(rdev);  		radeon_wb_fini(rdev);  		radeon_ib_pool_fini(rdev); @@ -2932,9 +3091,9 @@ int r600_init(struct radeon_device *rdev)  void r600_fini(struct radeon_device *rdev)  { +	radeon_pm_fini(rdev);  	r600_audio_fini(rdev);  	r600_cp_fini(rdev); -	r600_dma_fini(rdev);  	r600_irq_fini(rdev);  	radeon_wb_fini(rdev);  	radeon_ib_pool_fini(rdev); @@ -3350,7 +3509,6 @@ int r600_irq_set(struct radeon_device *rdev)  	u32 hpd1, hpd2, hpd3, hpd4 = 0, hpd5 = 0, hpd6 = 0;  	u32 grbm_int_cntl = 0;  	u32 hdmi0, hdmi1; -	u32 d1grph = 0, d2grph = 0;  	u32 dma_cntl;  	u32 thermal_int = 0; @@ -3459,8 +3617,8 @@ int r600_irq_set(struct radeon_device *rdev)  	WREG32(CP_INT_CNTL, cp_int_cntl);  	WREG32(DMA_CNTL, dma_cntl);  	WREG32(DxMODE_INT_MASK, mode_int); -	WREG32(D1GRPH_INTERRUPT_CONTROL, d1grph); -	WREG32(D2GRPH_INTERRUPT_CONTROL, d2grph); +	WREG32(D1GRPH_INTERRUPT_CONTROL, DxGRPH_PFLIP_INT_MASK); +	WREG32(D2GRPH_INTERRUPT_CONTROL, DxGRPH_PFLIP_INT_MASK);  	WREG32(GRBM_INT_CNTL, grbm_int_cntl);  	if (ASIC_IS_DCE3(rdev)) {  		WREG32(DC_HPD1_INT_CONTROL, hpd1); @@ -3637,6 +3795,7 @@ static u32 r600_get_ih_wptr(struct radeon_device *rdev)  		tmp = RREG32(IH_RB_CNTL);  		tmp |= IH_WPTR_OVERFLOW_CLEAR;  		WREG32(IH_RB_CNTL, tmp); +		wptr &= ~RB_OVERFLOW;  	}  	return (wptr & rdev->ih.ptr_mask);  } @@ -3721,7 +3880,7 @@ restart_ih:  						wake_up(&rdev->irq.vblank_queue);  					}  					if (atomic_read(&rdev->irq.pflip[0])) -						radeon_crtc_handle_flip(rdev, 0); +						radeon_crtc_handle_vblank(rdev, 0);  					rdev->irq.stat_regs.r600.disp_int &= ~LB_D1_VBLANK_INTERRUPT;  					DRM_DEBUG("IH: D1 vblank\n");  				} @@ -3747,7 +3906,7 @@ restart_ih:  						wake_up(&rdev->irq.vblank_queue);  					}  					if (atomic_read(&rdev->irq.pflip[1])) -						radeon_crtc_handle_flip(rdev, 1); +						radeon_crtc_handle_vblank(rdev, 1);  					rdev->irq.stat_regs.r600.disp_int &= ~LB_D2_VBLANK_INTERRUPT;  					DRM_DEBUG("IH: D2 vblank\n");  				} @@ -3763,6 +3922,14 @@ restart_ih:  				break;  			}  			break; +		case 9: /* D1 pflip */ +			DRM_DEBUG("IH: D1 flip\n"); +			radeon_crtc_handle_flip(rdev, 0); +			break; +		case 11: /* D2 pflip */ +			DRM_DEBUG("IH: D2 flip\n"); +			radeon_crtc_handle_flip(rdev, 1); +			break;  		case 19: /* HPD/DAC hotplug */  			switch (src_data) {  			case 0: @@ -3833,6 +4000,10 @@ restart_ih:  				break;  			}  			break; +		case 124: /* UVD */ +			DRM_DEBUG("IH: UVD int: 0x%08x\n", src_data); +			radeon_fence_process(rdev, R600_RING_TYPE_UVD_INDEX); +			break;  		case 176: /* CP_INT in ring buffer */  		case 177: /* CP_INT in IB1 */  		case 178: /* CP_INT in IB2 */  | 
