diff options
Diffstat (limited to 'drivers/video/fbdev/nvidia/nv_accel.c')
| -rw-r--r-- | drivers/video/fbdev/nvidia/nv_accel.c | 416 | 
1 files changed, 416 insertions, 0 deletions
diff --git a/drivers/video/fbdev/nvidia/nv_accel.c b/drivers/video/fbdev/nvidia/nv_accel.c new file mode 100644 index 00000000000..ad6472a894e --- /dev/null +++ b/drivers/video/fbdev/nvidia/nv_accel.c @@ -0,0 +1,416 @@ + /***************************************************************************\ +|*                                                                           *| +|*       Copyright 1993-2003 NVIDIA, Corporation.  All rights reserved.      *| +|*                                                                           *| +|*     NOTICE TO USER:   The source code  is copyrighted under  U.S. and     *| +|*     international laws.  Users and possessors of this source code are     *| +|*     hereby granted a nonexclusive,  royalty-free copyright license to     *| +|*     use this code in individual and commercial software.                  *| +|*                                                                           *| +|*     Any use of this source code must include,  in the user documenta-     *| +|*     tion and  internal comments to the code,  notices to the end user     *| +|*     as follows:                                                           *| +|*                                                                           *| +|*       Copyright 1993-2003 NVIDIA, Corporation.  All rights reserved.      *| +|*                                                                           *| +|*     NVIDIA, CORPORATION MAKES NO REPRESENTATION ABOUT THE SUITABILITY     *| +|*     OF  THIS SOURCE  CODE  FOR ANY PURPOSE.  IT IS  PROVIDED  "AS IS"     *| +|*     WITHOUT EXPRESS OR IMPLIED WARRANTY OF ANY KIND.  NVIDIA, CORPOR-     *| +|*     ATION DISCLAIMS ALL WARRANTIES  WITH REGARD  TO THIS SOURCE CODE,     *| +|*     INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY, NONINFRINGE-     *| +|*     MENT,  AND FITNESS  FOR A PARTICULAR PURPOSE.   IN NO EVENT SHALL     *| +|*     NVIDIA, CORPORATION  BE LIABLE FOR ANY SPECIAL,  INDIRECT,  INCI-     *| +|*     DENTAL, OR CONSEQUENTIAL DAMAGES,  OR ANY DAMAGES  WHATSOEVER RE-     *| +|*     SULTING FROM LOSS OF USE,  DATA OR PROFITS,  WHETHER IN AN ACTION     *| +|*     OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION,  ARISING OUT OF     *| +|*     OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOURCE CODE.     *| +|*                                                                           *| +|*     U.S. Government  End  Users.   This source code  is a "commercial     *| +|*     item,"  as that  term is  defined at  48 C.F.R. 2.101 (OCT 1995),     *| +|*     consisting  of "commercial  computer  software"  and  "commercial     *| +|*     computer  software  documentation,"  as such  terms  are  used in     *| +|*     48 C.F.R. 12.212 (SEPT 1995)  and is provided to the U.S. Govern-     *| +|*     ment only as  a commercial end item.   Consistent with  48 C.F.R.     *| +|*     12.212 and  48 C.F.R. 227.7202-1 through  227.7202-4 (JUNE 1995),     *| +|*     all U.S. Government End Users  acquire the source code  with only     *| +|*     those rights set forth herein.                                        *| +|*                                                                           *| + \***************************************************************************/ + +/* + * GPL Licensing Note - According to Mark Vojkovich, author of the Xorg/ + * XFree86 'nv' driver, this source code is provided under MIT-style licensing + * where the source code is provided "as is" without warranty of any kind. + * The only usage restriction is for the copyright notices to be retained + * whenever code is used. + * + * Antonino Daplas <adaplas@pol.net> 2005-03-11 + */ + +#include <linux/fb.h> +#include "nv_type.h" +#include "nv_proto.h" +#include "nv_dma.h" +#include "nv_local.h" + +/* There is a HW race condition with videoram command buffers. +   You can't jump to the location of your put offset.  We write put +   at the jump offset + SKIPS dwords with noop padding in between +   to solve this problem */ +#define SKIPS  8 + +static const int NVCopyROP[16] = { +	0xCC,			/* copy   */ +	0x55			/* invert */ +}; + +static const int NVCopyROP_PM[16] = { +	0xCA,			/* copy  */ +	0x5A,			/* invert */ +}; + +static inline void nvidiafb_safe_mode(struct fb_info *info) +{ +	struct nvidia_par *par = info->par; + +	touch_softlockup_watchdog(); +	info->pixmap.scan_align = 1; +	par->lockup = 1; +} + +static inline void NVFlush(struct fb_info *info) +{ +	struct nvidia_par *par = info->par; +	int count = 1000000000; + +	while (--count && READ_GET(par) != par->dmaPut) ; + +	if (!count) { +		printk("nvidiafb: DMA Flush lockup\n"); +		nvidiafb_safe_mode(info); +	} +} + +static inline void NVSync(struct fb_info *info) +{ +	struct nvidia_par *par = info->par; +	int count = 1000000000; + +	while (--count && NV_RD32(par->PGRAPH, 0x0700)) ; + +	if (!count) { +		printk("nvidiafb: DMA Sync lockup\n"); +		nvidiafb_safe_mode(info); +	} +} + +static void NVDmaKickoff(struct nvidia_par *par) +{ +	if (par->dmaCurrent != par->dmaPut) { +		par->dmaPut = par->dmaCurrent; +		WRITE_PUT(par, par->dmaPut); +	} +} + +static void NVDmaWait(struct fb_info *info, int size) +{ +	struct nvidia_par *par = info->par; +	int dmaGet; +	int count = 1000000000, cnt; +	size++; + +	while (par->dmaFree < size && --count && !par->lockup) { +		dmaGet = READ_GET(par); + +		if (par->dmaPut >= dmaGet) { +			par->dmaFree = par->dmaMax - par->dmaCurrent; +			if (par->dmaFree < size) { +				NVDmaNext(par, 0x20000000); +				if (dmaGet <= SKIPS) { +					if (par->dmaPut <= SKIPS) +						WRITE_PUT(par, SKIPS + 1); +					cnt = 1000000000; +					do { +						dmaGet = READ_GET(par); +					} while (--cnt && dmaGet <= SKIPS); +					if (!cnt) { +						printk("DMA Get lockup\n"); +						par->lockup = 1; +					} +				} +				WRITE_PUT(par, SKIPS); +				par->dmaCurrent = par->dmaPut = SKIPS; +				par->dmaFree = dmaGet - (SKIPS + 1); +			} +		} else +			par->dmaFree = dmaGet - par->dmaCurrent - 1; +	} + +	if (!count) { +		printk("nvidiafb: DMA Wait Lockup\n"); +		nvidiafb_safe_mode(info); +	} +} + +static void NVSetPattern(struct fb_info *info, u32 clr0, u32 clr1, +			 u32 pat0, u32 pat1) +{ +	struct nvidia_par *par = info->par; + +	NVDmaStart(info, par, PATTERN_COLOR_0, 4); +	NVDmaNext(par, clr0); +	NVDmaNext(par, clr1); +	NVDmaNext(par, pat0); +	NVDmaNext(par, pat1); +} + +static void NVSetRopSolid(struct fb_info *info, u32 rop, u32 planemask) +{ +	struct nvidia_par *par = info->par; + +	if (planemask != ~0) { +		NVSetPattern(info, 0, planemask, ~0, ~0); +		if (par->currentRop != (rop + 32)) { +			NVDmaStart(info, par, ROP_SET, 1); +			NVDmaNext(par, NVCopyROP_PM[rop]); +			par->currentRop = rop + 32; +		} +	} else if (par->currentRop != rop) { +		if (par->currentRop >= 16) +			NVSetPattern(info, ~0, ~0, ~0, ~0); +		NVDmaStart(info, par, ROP_SET, 1); +		NVDmaNext(par, NVCopyROP[rop]); +		par->currentRop = rop; +	} +} + +static void NVSetClippingRectangle(struct fb_info *info, int x1, int y1, +				   int x2, int y2) +{ +	struct nvidia_par *par = info->par; +	int h = y2 - y1 + 1; +	int w = x2 - x1 + 1; + +	NVDmaStart(info, par, CLIP_POINT, 2); +	NVDmaNext(par, (y1 << 16) | x1); +	NVDmaNext(par, (h << 16) | w); +} + +void NVResetGraphics(struct fb_info *info) +{ +	struct nvidia_par *par = info->par; +	u32 surfaceFormat, patternFormat, rectFormat, lineFormat; +	int pitch, i; + +	pitch = info->fix.line_length; + +	par->dmaBase = (u32 __iomem *) (&par->FbStart[par->FbUsableSize]); + +	for (i = 0; i < SKIPS; i++) +		NV_WR32(&par->dmaBase[i], 0, 0x00000000); + +	NV_WR32(&par->dmaBase[0x0 + SKIPS], 0, 0x00040000); +	NV_WR32(&par->dmaBase[0x1 + SKIPS], 0, 0x80000010); +	NV_WR32(&par->dmaBase[0x2 + SKIPS], 0, 0x00042000); +	NV_WR32(&par->dmaBase[0x3 + SKIPS], 0, 0x80000011); +	NV_WR32(&par->dmaBase[0x4 + SKIPS], 0, 0x00044000); +	NV_WR32(&par->dmaBase[0x5 + SKIPS], 0, 0x80000012); +	NV_WR32(&par->dmaBase[0x6 + SKIPS], 0, 0x00046000); +	NV_WR32(&par->dmaBase[0x7 + SKIPS], 0, 0x80000013); +	NV_WR32(&par->dmaBase[0x8 + SKIPS], 0, 0x00048000); +	NV_WR32(&par->dmaBase[0x9 + SKIPS], 0, 0x80000014); +	NV_WR32(&par->dmaBase[0xA + SKIPS], 0, 0x0004A000); +	NV_WR32(&par->dmaBase[0xB + SKIPS], 0, 0x80000015); +	NV_WR32(&par->dmaBase[0xC + SKIPS], 0, 0x0004C000); +	NV_WR32(&par->dmaBase[0xD + SKIPS], 0, 0x80000016); +	NV_WR32(&par->dmaBase[0xE + SKIPS], 0, 0x0004E000); +	NV_WR32(&par->dmaBase[0xF + SKIPS], 0, 0x80000017); + +	par->dmaPut = 0; +	par->dmaCurrent = 16 + SKIPS; +	par->dmaMax = 8191; +	par->dmaFree = par->dmaMax - par->dmaCurrent; + +	switch (info->var.bits_per_pixel) { +	case 32: +	case 24: +		surfaceFormat = SURFACE_FORMAT_DEPTH24; +		patternFormat = PATTERN_FORMAT_DEPTH24; +		rectFormat = RECT_FORMAT_DEPTH24; +		lineFormat = LINE_FORMAT_DEPTH24; +		break; +	case 16: +		surfaceFormat = SURFACE_FORMAT_DEPTH16; +		patternFormat = PATTERN_FORMAT_DEPTH16; +		rectFormat = RECT_FORMAT_DEPTH16; +		lineFormat = LINE_FORMAT_DEPTH16; +		break; +	default: +		surfaceFormat = SURFACE_FORMAT_DEPTH8; +		patternFormat = PATTERN_FORMAT_DEPTH8; +		rectFormat = RECT_FORMAT_DEPTH8; +		lineFormat = LINE_FORMAT_DEPTH8; +		break; +	} + +	NVDmaStart(info, par, SURFACE_FORMAT, 4); +	NVDmaNext(par, surfaceFormat); +	NVDmaNext(par, pitch | (pitch << 16)); +	NVDmaNext(par, 0); +	NVDmaNext(par, 0); + +	NVDmaStart(info, par, PATTERN_FORMAT, 1); +	NVDmaNext(par, patternFormat); + +	NVDmaStart(info, par, RECT_FORMAT, 1); +	NVDmaNext(par, rectFormat); + +	NVDmaStart(info, par, LINE_FORMAT, 1); +	NVDmaNext(par, lineFormat); + +	par->currentRop = ~0;	/* set to something invalid */ +	NVSetRopSolid(info, ROP_COPY, ~0); + +	NVSetClippingRectangle(info, 0, 0, info->var.xres_virtual, +			       info->var.yres_virtual); + +	NVDmaKickoff(par); +} + +int nvidiafb_sync(struct fb_info *info) +{ +	struct nvidia_par *par = info->par; + +	if (info->state != FBINFO_STATE_RUNNING) +		return 0; + +	if (!par->lockup) +		NVFlush(info); + +	if (!par->lockup) +		NVSync(info); + +	return 0; +} + +void nvidiafb_copyarea(struct fb_info *info, const struct fb_copyarea *region) +{ +	struct nvidia_par *par = info->par; + +	if (info->state != FBINFO_STATE_RUNNING) +		return; + +	if (par->lockup) { +		cfb_copyarea(info, region); +		return; +	} + +	NVDmaStart(info, par, BLIT_POINT_SRC, 3); +	NVDmaNext(par, (region->sy << 16) | region->sx); +	NVDmaNext(par, (region->dy << 16) | region->dx); +	NVDmaNext(par, (region->height << 16) | region->width); + +	NVDmaKickoff(par); +} + +void nvidiafb_fillrect(struct fb_info *info, const struct fb_fillrect *rect) +{ +	struct nvidia_par *par = info->par; +	u32 color; + +	if (info->state != FBINFO_STATE_RUNNING) +		return; + +	if (par->lockup) { +		cfb_fillrect(info, rect); +		return; +	} + +	if (info->var.bits_per_pixel == 8) +		color = rect->color; +	else +		color = ((u32 *) info->pseudo_palette)[rect->color]; + +	if (rect->rop != ROP_COPY) +		NVSetRopSolid(info, rect->rop, ~0); + +	NVDmaStart(info, par, RECT_SOLID_COLOR, 1); +	NVDmaNext(par, color); + +	NVDmaStart(info, par, RECT_SOLID_RECTS(0), 2); +	NVDmaNext(par, (rect->dx << 16) | rect->dy); +	NVDmaNext(par, (rect->width << 16) | rect->height); + +	NVDmaKickoff(par); + +	if (rect->rop != ROP_COPY) +		NVSetRopSolid(info, ROP_COPY, ~0); +} + +static void nvidiafb_mono_color_expand(struct fb_info *info, +				       const struct fb_image *image) +{ +	struct nvidia_par *par = info->par; +	u32 fg, bg, mask = ~(~0 >> (32 - info->var.bits_per_pixel)); +	u32 dsize, width, *data = (u32 *) image->data, tmp; +	int j, k = 0; + +	width = (image->width + 31) & ~31; +	dsize = (width * image->height) >> 5; + +	if (info->var.bits_per_pixel == 8) { +		fg = image->fg_color | mask; +		bg = image->bg_color | mask; +	} else { +		fg = ((u32 *) info->pseudo_palette)[image->fg_color] | mask; +		bg = ((u32 *) info->pseudo_palette)[image->bg_color] | mask; +	} + +	NVDmaStart(info, par, RECT_EXPAND_TWO_COLOR_CLIP, 7); +	NVDmaNext(par, (image->dy << 16) | (image->dx & 0xffff)); +	NVDmaNext(par, ((image->dy + image->height) << 16) | +		  ((image->dx + image->width) & 0xffff)); +	NVDmaNext(par, bg); +	NVDmaNext(par, fg); +	NVDmaNext(par, (image->height << 16) | width); +	NVDmaNext(par, (image->height << 16) | width); +	NVDmaNext(par, (image->dy << 16) | (image->dx & 0xffff)); + +	while (dsize >= RECT_EXPAND_TWO_COLOR_DATA_MAX_DWORDS) { +		NVDmaStart(info, par, RECT_EXPAND_TWO_COLOR_DATA(0), +			   RECT_EXPAND_TWO_COLOR_DATA_MAX_DWORDS); + +		for (j = RECT_EXPAND_TWO_COLOR_DATA_MAX_DWORDS; j--;) { +			tmp = data[k++]; +			reverse_order(&tmp); +			NVDmaNext(par, tmp); +		} + +		dsize -= RECT_EXPAND_TWO_COLOR_DATA_MAX_DWORDS; +	} + +	if (dsize) { +		NVDmaStart(info, par, RECT_EXPAND_TWO_COLOR_DATA(0), dsize); + +		for (j = dsize; j--;) { +			tmp = data[k++]; +			reverse_order(&tmp); +			NVDmaNext(par, tmp); +		} +	} + +	NVDmaKickoff(par); +} + +void nvidiafb_imageblit(struct fb_info *info, const struct fb_image *image) +{ +	struct nvidia_par *par = info->par; + +	if (info->state != FBINFO_STATE_RUNNING) +		return; + +	if (image->depth == 1 && !par->lockup) +		nvidiafb_mono_color_expand(info, image); +	else +		cfb_imageblit(info, image); +}  | 
