Shading of images using texture

Computer graphics processing and selective visual display system – Computer graphics processing – Attributes

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

C345S426000

Reexamination Certificate

active

07098925

ABSTRACT:
Colors are obtained and used to shade two-dimensional non-photo realistic images of three-dimensional models. The colors are stored as texels in texture maps. The texels are chosen to shade the images based on their orientation in a scene. The colors may be obtained once, in a pre-rendering process. The colors may then be selected in a run-time process.

REFERENCES:
patent: 4600919 (1986-07-01), Stern
patent: 5124914 (1992-06-01), Grangeat
patent: 5163126 (1992-11-01), Einkauf et al.
patent: 5490240 (1996-02-01), Foran et al.
patent: 5659672 (1997-08-01), Yutaka et al.
patent: 5731819 (1998-03-01), Gagne et al.
patent: 5742292 (1998-04-01), Murata
patent: 5764228 (1998-06-01), Baldwin
patent: 5856829 (1999-01-01), Gray, III et al.
patent: 5870097 (1999-02-01), Snyder et al.
patent: 5905503 (1999-05-01), Penna
patent: 5966133 (1999-10-01), Hoppe
patent: 6037949 (2000-03-01), DeRose et al.
patent: 6057850 (2000-05-01), Kichury
patent: 6057859 (2000-05-01), Handelman et al.
patent: 6064393 (2000-05-01), Lengyel et al.
patent: 6078332 (2000-06-01), Ohazama
patent: 6163319 (2000-12-01), Peercy et al.
patent: 6208347 (2001-03-01), Migdal et al.
patent: 6226005 (2001-05-01), Laferriere
patent: 6252608 (2001-06-01), Snyder et al.
patent: 6313846 (2001-11-01), Fenney et al.
patent: 6320578 (2001-11-01), Shiitani et al.
patent: 6337880 (2002-01-01), Cornog et al.
patent: 6339430 (2002-01-01), Takeshita
patent: 6388670 (2002-05-01), Naka et al.
patent: 6771264 (2004-08-01), Duluk et al.
Lewis “Pose Space Deformation: A Unified Approach to Shape Interpolation and Skeleton-Driven Deformation” Centropolis, New Orleans, LA, 165-172.
Lasseter “Principles of Traditional Animation Applied to 3D Computer Animation” Pixar, San Rafael, California, 1987.
Thomas (Contributor) et al., “The Illusion of Life: Disney Animation” 47-51.
Hoppe, “Progressive Meshes” Microsoft Research, 99-108, http://www.research.microsft.com/research/graphics/hoppe/.
Popovic et al., “Progressive Simplicial Complexes” Microsoft Research, http://www.research.microsft.com/˜hoppe/.
Hoppe “Efficient Implementation of progressive meshes” Coput. & Graphics vol. 22, No. 1, pp. 27-36, 1998.
Taubin et al., “Progressive Forest Spilt Compression” IBM T.J. Watson Research Center, Yorktown Heights, NY.
Cohen-Or et al., “Progressive Compression of Arbitrary Triangular Meshes” Computer Science Department, School of Mathematical Sciences, Tel Aviv, Israel.
Bajaj et al., “Progressive Compression and Transmission of Arbitrary Triangular Meshes” Department of Computer Sciences, University of Texas at Austin, Austin, TX.
Pajarola et al., “Compressed Progressive Meshes” Graphics, Visualization & Usability Center, College of Computing, Georgia Institute of Technology, Jan. 1999.
Alliez et al., “Progressive Compression for Lossless Transmission of Triangle Meshes” University of Southern California, Los Angeles, CA, 195-202.
Chow “Optimized Geometry Compression for Real-time Rendering” Massachusetts Institute of Technology, Proceedings Visualization 1997, Oct. 19-24, 1997, Phoenix, AZ, 347-354.
Markosian “Real-Time Nonphotorealistic Rendering” Brown University site of the NSF Science and Technology Center for Computer Graphics and Scientific Visualization, Providence, RI.
Elber Line Art Rendering via a Coverage of Isoperimetric Curves, IEEE Transactions on Visualization and Computer Graphics, vol. 1, Department of Computer Science, Technion, Israel Institute of Technology, Haifa, Israel, Sep. 1995.
Zeleznik et al., “SKETCH: An Interface for Sketching 3D Scenes” Brown University site of the NSF Science and Technology Center for Computer Graphics and Scientific Visualization, 1996.
Landsdown et al., “Expressive Rendering: A Review of Nonphotorealistic Techniques” IEEE Computer graphics and Applications, 29-37, 1995.
Raskar “Image Precision Silhouette Edges” University of North Carolina at Chapel Hill, Microsoft Research, 1999 Symposium on Interactive 3D Graphics Atlanta, GA, 135-231, 1999.
Ma et al., “Extracting Feature Lines for 3D Unstructured Grids” Institute for Computer Applications in Science and Engineering (ICASE), NASA Langley Research Center, Hampton, VA, IEEE, 1997.
Samet “Applications of spatial data structures: computer graphics, image processing, and GIS” University of Maryland, Addison-Wesley Publishing Company, 1060-1064, Reading, MA, Jun. 1990.
Dyn “A Butterfly Subdivision Scheme for Surface Interpolation with Tension Control” ACM Transactions on Graphics, vol. 9, No. 2, Apr. 1990.
Zorin “Interpolation Subdivision for Meshes With Arbitrary Topology” Department of Computer Science, California Institute of Technology, Pasadena, CA.
Lee “Navigating through Triangle Meshes Implemented as linear Quadtrees” Computer Science Department, Center for Automation Research, Institute for Advanced Computer Studies, University of Maryland College Park, MD, Apr. 1998.
Grooch et al., “Interactive Technical Illustration,” Proceedings 1999 Symposium on Interactive 3D Graphics, Apr. 26-28, 1999, pp 31-38 and p 220.
Heidrich et al., “Realistic, Hardware-Accelerated Shading and Lighting,” Computer Graphics Proceedings, Annual Conference Series, 1999, pp 171-178.

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Shading of images using texture does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Shading of images using texture, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Shading of images using texture will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-3698081

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.