SDWebImage主线设计的解码大致有两种:普通解码和渐进式解码。本文只对普通解码进行解析。
普通解码又分为正常解码和大图解码。
普通解码
普通解码从 -[SDWebImageDownloaderOperation URLSession:task:didCompleteWithError:]
发起
更多解析请参考SDWebImage主线梳理(二)
同步锁加持,发送两个通知:
SDWebImageDownloadStopNotification
,SDWebImageDownloadFinishNotification
如果 self.callbackBlocks 有 LoaderCompletedBlock(key=kCompletedCallbackKey), 继续
self.imageData(NSMutableData) 在
didReceiveData:
方法中一点一点拼接的可变data异步串行队列(self.coderQueue),调用 SDImageLoaderDecodeImageData() 解码imageData,输出UIImage
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
解码函数 SDImageLoaderDecodeImageData()
SDImageLoaderDecodeImageData()
函数在 SDImageLoader.m
中
-
SDImageScaleFactorForKey(NSString * key),返回一个屏幕倍数因子。
- 如果key是普通名字,判断key包不包含"@2x."或者"@3x.",包含就返回这个倍数因子。
- 如果key是一个URL,百分号编码的下的@ = %40,判断key包不包含"%402x."或者"%403x."。
如果不是“仅解码第一帧”并且是动图,则装载动图的所有帧
如果不是动图,
-[SDImageCodersManager(sharedManager) decodedImageWithData:imageData options:]
,参考详解一动图不解码
如果应该解码,判断是否应该按比例缩小图片
按比例缩小图片:
-[SDImageCoderHelper decodedAndScaledDownImageWithImage:limitBytes:]
,参考详解二不缩小图片:
-[SDImageCoderHelper decodedImageWithImage:]
,参考详解三
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
详解一
+[SDImageCodersManager sharedManager] 方法
调用 -[SDImageCodersManager new]
即 -[SDImageCodersManager init]
+ (nonnull instancetype)sharedManager {
static dispatch_once_t once;
static id instance;
dispatch_once(&once, ^{
instance = [self new];
});
return instance;
}
-[SDImageCodersManager init] 方法
- 创建可变数组,初始化赋值 SDImageIOCoder, SDImageGIFCoder, SDImageAPNGCoder; 都是调用 sharedCoder 方法获取, 待展开
- 创建信号量线程锁,保存在 SDImageCodersManager 的 _codersLock 属性中
- (instancetype)init {
if (self = [super init]) {
// initialize with default coders
_imageCoders = [NSMutableArray arrayWithArray:@[[SDImageIOCoder sharedCoder], [SDImageGIFCoder sharedCoder], [SDImageAPNGCoder sharedCoder]]];
_codersLock = dispatch_semaphore_create(1);
}
return self;
}
-[SDImageCodersManager decodedImageWithData:options:] 方法
取出(加锁) _imageCoders, SDImageIOCoder, SDImageGIFCoder, SDImageAPNGCoder;
反转遍历 _imageCoders ,顺序为 SDImageAPNGCoder -> SDImageGIFCoder -> SDImageIOCoder
-
每个 coder 都调用
canDecodeFromData:
方法,判断是否是可以解码的格式-
-[SDImageIOAnimatedCoder canDecodeFromData:]
,取两个值(图片格式)进行对比-
+[NSData(ImageContentType) sd_imageFormatForImageData:]
获取data中包含的图片格式-
-[NSData getBytes:ength:]
获取的一个字节(第一个字节)可以区分图片格式
-
- (self.class.imageFormat 也就是
+[SDImageAPNGCoder imageFormat]
) == SDImageFormatPNG
-
-
-
如果返回值为YES则 coder 调用
decodedImageWithData:options:
方法,输出 image。-
-[SDImageIOAnimatedCoder decodedImageWithData:options:]
::: 此处并不是解码的地方,只是将压缩的图片二进制流读取到UIImage中。
-
打断for循环,返回image
// -[SDImageCodersManager decodedImageWithData:options:]
- (UIImage *)decodedImageWithData:(NSData *)data options:(nullable SDImageCoderOptions *)options {
if (!data) {
return nil;
}
UIImage *image;
NSArray> *coders = self.coders;
for (id coder in coders.reverseObjectEnumerator) {
if ([coder canDecodeFromData:data]) {
image = [coder decodedImageWithData:data options:options];
break;
}
}
return image;
}
// -[SDImageIOAnimatedCoder canDecodeFromData:]
- (BOOL)canDecodeFromData:(nullable NSData *)data {
return ([NSData sd_imageFormatForImageData:data] == self.class.imageFormat);
}
@implementation NSData (ImageContentType)
+ (SDImageFormat)sd_imageFormatForImageData:(nullable NSData *)data {
if (!data) {
return SDImageFormatUndefined;
}
// File signatures table: http://www.garykessler.net/library/file_sigs.html
uint8_t c;
[data getBytes:&c length:1];
switch (c) {
case 0xFF:
return SDImageFormatJPEG;
case 0x89:
return SDImageFormatPNG;
case 0x47:
return SDImageFormatGIF;
case 0x49:
case 0x4D:
return SDImageFormatTIFF;
case 0x52: {
if (data.length >= 12) {
//RIFF....WEBP
NSString *testString = [[NSString alloc] initWithData:[data subdataWithRange:NSMakeRange(0, 12)] encoding:NSASCIIStringEncoding];
if ([testString hasPrefix:@"RIFF"] && [testString hasSuffix:@"WEBP"]) {
return SDImageFormatWebP;
}
}
break;
}
case 0x00: {
if (data.length >= 12) {
//....ftypheic ....ftypheix ....ftyphevc ....ftyphevx
NSString *testString = [[NSString alloc] initWithData:[data subdataWithRange:NSMakeRange(4, 8)] encoding:NSASCIIStringEncoding];
if ([testString isEqualToString:@"ftypheic"]
|| [testString isEqualToString:@"ftypheix"]
|| [testString isEqualToString:@"ftyphevc"]
|| [testString isEqualToString:@"ftyphevx"]) {
return SDImageFormatHEIC;
}
//....ftypmif1 ....ftypmsf1
if ([testString isEqualToString:@"ftypmif1"] || [testString isEqualToString:@"ftypmsf1"]) {
return SDImageFormatHEIF;
}
}
break;
}
}
return SDImageFormatUndefined;
}
// -[SDImageIOAnimatedCoder decodedImageWithData:options:]
- (UIImage *)decodedImageWithData:(NSData *)data options:(nullable SDImageCoderOptions *)options {
if (!data) {
return nil;
}
CGFloat scale = 1;
NSNumber *scaleFactor = options[SDImageCoderDecodeScaleFactor];
if (scaleFactor != nil) {
scale = MAX([scaleFactor doubleValue], 1);
}
#if SD_MAC
SDAnimatedImageRep *imageRep = [[SDAnimatedImageRep alloc] initWithData:data];
NSSize size = NSMakeSize(imageRep.pixelsWide / scale, imageRep.pixelsHigh / scale);
imageRep.size = size;
NSImage *animatedImage = [[NSImage alloc] initWithSize:size];
[animatedImage addRepresentation:imageRep];
return animatedImage;
#else
CGImageSourceRef source = CGImageSourceCreateWithData((__bridge CFDataRef)data, NULL);
if (!source) {
return nil;
}
size_t count = CGImageSourceGetCount(source);
UIImage *animatedImage;
BOOL decodeFirstFrame = [options[SDImageCoderDecodeFirstFrameOnly] boolValue];
if (decodeFirstFrame || count <= 1) {
animatedImage = [[UIImage alloc] initWithData:data scale:scale];
} else {
NSMutableArray *frames = [NSMutableArray array];
for (size_t i = 0; i < count; i++) {
CGImageRef imageRef = CGImageSourceCreateImageAtIndex(source, i, NULL);
if (!imageRef) {
continue;
}
NSTimeInterval duration = [self.class frameDurationAtIndex:i source:source];
UIImage *image = [[UIImage alloc] initWithCGImage:imageRef scale:scale orientation:UIImageOrientationUp];
CGImageRelease(imageRef);
SDImageFrame *frame = [SDImageFrame frameWithImage:image duration:duration];
[frames addObject:frame];
}
NSUInteger loopCount = [self.class imageLoopCountWithSource:source];
animatedImage = [SDImageCoderHelper animatedImageWithFrames:frames];
animatedImage.sd_imageLoopCount = loopCount;
}
animatedImage.sd_imageFormat = self.class.imageFormat;
CFRelease(source);
return animatedImage;
#endif
}
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
详解二-解码大图
+[SDImageCoderHelper decodedAndScaledDownImageWithImage:limitBytes:] 方法
-
+[SDImageCoderHelper shouldDecodeImage:]
- 已解码过(判断关联属性sd_isDecoded)的不再解码;
- 图片为nil不解码;
- 动图不解码;
-
+[SDImageCoderHelper shouldScaleDownImage:limitBytes:]
,判断是否是要比例缩小,如果不需要则直接正常解码图片即可- 限制图片当超过多少字节时需要缩小,可以手动设置也可以走SD的默认值
- SD默认当图片总像素数量超过 262144(60M所拥有的像素数量)需要缩小
- 只要限制的像素数量比源像素数量小即需要缩小。PS:太小了也不行,至少要超过(大于) 1M 才需要。
初始化目标总像素量(destTotalPixels,60M的总像素) 以及目标图片的单片(tile)总像素量(tileTotalPixels,20M的总像素)
更多的解析看源码中添加的注释吧
+ (UIImage *)decodedAndScaledDownImageWithImage:(UIImage *)image limitBytes:(NSUInteger)bytes {
#if SD_MAC
return image;
#else
if (![self shouldDecodeImage:image]) {
return image;
}
if (![self shouldScaleDownImage:image limitBytes:bytes]) {
return [self decodedImageWithImage:image];
}
CGFloat destTotalPixels;
CGFloat tileTotalPixels;
if (bytes > 0) {
destTotalPixels = bytes / kBytesPerPixel;
tileTotalPixels = destTotalPixels / 3;
} else {
destTotalPixels = kDestTotalPixels;
tileTotalPixels = kTileTotalPixels;
}
CGContextRef destContext;
// autorelease the bitmap context and all vars to help system to free memory when there are memory warning.
// on iOS7, do not forget to call [[SDImageCache sharedImageCache] clearMemory];
@autoreleasepool {
CGImageRef sourceImageRef = image.CGImage;
CGSize sourceResolution = CGSizeZero;
sourceResolution.width = CGImageGetWidth(sourceImageRef);
sourceResolution.height = CGImageGetHeight(sourceImageRef);
CGFloat sourceTotalPixels = sourceResolution.width * sourceResolution.height;
// Determine the scale ratio to apply to the input image
// that results in an output image of the defined size.
// see kDestImageSizeMB, and how it relates to destTotalPixels.
// 此处开方,假设destTotalPixels = 60M(拥有的像素量), sourceTotalPixels = 240M, imageScale = 1/2。
// 因为dest的总像素量需要宽高相乘才能得到,所以需要比例值需要开方。
CGFloat imageScale = sqrt(destTotalPixels / sourceTotalPixels);
CGSize destResolution = CGSizeZero;
destResolution.width = (int)(sourceResolution.width * imageScale);
destResolution.height = (int)(sourceResolution.height * imageScale);
// device color space
CGColorSpaceRef colorspaceRef = [self colorSpaceGetDeviceRGB];
BOOL hasAlpha = [self CGImageContainsAlpha:sourceImageRef];
// iOS display alpha info (BGRA8888/BGRX8888)
CGBitmapInfo bitmapInfo = kCGBitmapByteOrder32Host;
bitmapInfo |= hasAlpha ? kCGImageAlphaPremultipliedFirst : kCGImageAlphaNoneSkipFirst;
// kCGImageAlphaNone is not supported in CGBitmapContextCreate.
// Since the original image here has no alpha info, use kCGImageAlphaNoneSkipFirst
// to create bitmap graphics contexts without alpha info.
destContext = CGBitmapContextCreate(NULL,
destResolution.width,
destResolution.height,
kBitsPerComponent,
0,
colorspaceRef,
bitmapInfo);
if (destContext == NULL) {
return image;
}
CGContextSetInterpolationQuality(destContext, kCGInterpolationHigh);
// Now define the size of the rectangle to be used for the
// incremental blits from the input image to the output image.
// we use a source tile width equal to the width of the source
// image due to the way that iOS retrieves image data from disk.
// iOS must decode an image from disk in full width 'bands', even
// if current graphics context is clipped to a subrect within that
// band. Therefore we fully utilize all of the pixel data that results
// from a decoding opertion by achnoring our tile size to the full
// width of the input image.
CGRect sourceTile = CGRectZero;
sourceTile.size.width = sourceResolution.width;
// The source tile height is dynamic. Since we specified the size
// of the source tile in MB, see how many rows of pixels high it
// can be given the input image width.
sourceTile.size.height = (int)(tileTotalPixels / sourceTile.size.width );
sourceTile.origin.x = 0.0f;
// The output tile is the same proportions as the input tile, but
// scaled to image scale.
CGRect destTile;
destTile.size.width = destResolution.width;
destTile.size.height = sourceTile.size.height * imageScale;
destTile.origin.x = 0.0f;
// The source seem overlap is proportionate to the destination seem overlap.
// this is the amount of pixels to overlap each tile as we assemble the ouput image.
float sourceSeemOverlap = (int)((kDestSeemOverlap/destResolution.height)*sourceResolution.height);
CGImageRef sourceTileImageRef;
// calculate the number of read/write operations required to assemble the
// output image.
int iterations = (int)( sourceResolution.height / sourceTile.size.height );
// If tile height doesn't divide the image height evenly, add another iteration
// to account for the remaining pixels.
int remainder = (int)sourceResolution.height % (int)sourceTile.size.height;
if(remainder) {
iterations++;
}
// Add seem overlaps to the tiles, but save the original tile height for y coordinate calculations.
float sourceTileHeightMinusOverlap = sourceTile.size.height;
sourceTile.size.height += sourceSeemOverlap;
destTile.size.height += kDestSeemOverlap;
for( int y = 0; y < iterations; ++y ) {
@autoreleasepool {
sourceTile.origin.y = y * sourceTileHeightMinusOverlap + sourceSeemOverlap;
destTile.origin.y = destResolution.height - (( y + 1 ) * sourceTileHeightMinusOverlap * imageScale + kDestSeemOverlap);// 用户空间和设备空间的坐标系转换
sourceTileImageRef = CGImageCreateWithImageInRect( sourceImageRef, sourceTile );
if( y == iterations - 1 && remainder ) {// 最后一次并且高不能被整除
float dify = destTile.size.height;
// 获取不能被整除的最后一小块真实高度
destTile.size.height = CGImageGetHeight( sourceTileImageRef ) * imageScale;
// 修正y值
dify -= destTile.size.height;
destTile.origin.y += dify;
}
CGContextDrawImage( destContext, destTile, sourceTileImageRef );
CGImageRelease( sourceTileImageRef );
}
}
CGImageRef destImageRef = CGBitmapContextCreateImage(destContext);
CGContextRelease(destContext);
if (destImageRef == NULL) {
return image;
}
UIImage *destImage = [[UIImage alloc] initWithCGImage:destImageRef scale:image.scale orientation:image.imageOrientation];
CGImageRelease(destImageRef);
if (destImage == nil) {
return image;
}
destImage.sd_isDecoded = YES;
destImage.sd_imageFormat = image.sd_imageFormat;
return destImage;
}
#endif
}
这个解码大图不是SD的原创,是苹果官方给出的解决方案,提供demo可下载。想要了解更多,就去《Large Image Downsizing》探索伟大航路吧!
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
详解三-正常解码
+[SDImageCoderHelper decodedImageWithImage:] 方法
+[SDImageCoderHelper shouldDecodeImage:]:已解码过(判断关联属性sd_isDecoded)的不再解码;图片为nil不解码;动图不解码;
+[SDImageCoderHelper CGImageCreateDecoded:]
-
+[SDImageCoderHelper CGImageCreateDecoded:orientation:]:输入image.CGImage,输出解码的CGImageRef; 核心解码函数:CGContextDrawImage()
获取图片的宽高,如果方向是左、右旋转则交换宽高数据
判断是否含有alpha信息
获取32位字节顺序(kCGBitmapByteOrder32Host这个宏避免考虑大小端问题),保存到位图信息bitmapInfo
bitmapInfo 按位或添加像素格式(alpha信息)。有alpha选择kCGImageAlphaPremultipliedFirst(BGRA8888), 无alpha选择kCGImageAlphaNoneSkipFirst
调用 +[SDImageCoderHelper colorSpaceGetDeviceRGB] 获取颜色空间,一个单例
调用 CGBitmapContextCreate() 函数创建位图上下文
调用 SDCGContextTransformFromOrientation() 获取方向旋转的 CGAffineTransform,调用 CGContextConcatCTM 关联到位图上下文,坐标系转换
-
❗️解码:CGContextDrawImage()❗️
- 传入的 bytesPerRow 参数是 0,正常情况下应传入 width * bytesPerPixel,但是这里传入 0 系统会帮助计算,而且系统还加了点小优化。
- 可以明显的看到Xcode检测到的内存情况和CPU使用情况的变化
获取位图:CGBitmapContextCreateImage(),输入context,输出解码后的CGImageRef。
释放上下文context
-[UIImage initWithCGImage:scale:orientation:],CGImage 转为 UIImage
UIImage 关联属性赋值:decodedImage.sd_isDecoded = YES; decodedImage.sd_imageFormat = image.sd_imageFormat;
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
另外再提一句
函数 SDImageCacheDecodeImageData() 也是普通的正常解码,与我们前面说的解码函数 SDImageLoaderDecodeImageData() 逻辑基本一致。
函数实现在 SDImageCacheDefine.m
文件中,此文件仅仅只写了这一个函数,其它什么都没有。
UIImage * _Nullable SDImageCacheDecodeImageData(NSData * _Nonnull imageData, NSString * _Nonnull cacheKey, SDWebImageOptions options, SDWebImageContext * _Nullable context) {
UIImage *image;
BOOL decodeFirstFrame = SD_OPTIONS_CONTAINS(options, SDWebImageDecodeFirstFrameOnly);
NSNumber *scaleValue = context[SDWebImageContextImageScaleFactor];
CGFloat scale = scaleValue.doubleValue >= 1 ? scaleValue.doubleValue : SDImageScaleFactorForKey(cacheKey);
SDImageCoderOptions *coderOptions = @{SDImageCoderDecodeFirstFrameOnly : @(decodeFirstFrame), SDImageCoderDecodeScaleFactor : @(scale)};
if (context) {
SDImageCoderMutableOptions *mutableCoderOptions = [coderOptions mutableCopy];
[mutableCoderOptions setValue:context forKey:SDImageCoderWebImageContext];
coderOptions = [mutableCoderOptions copy];
}
if (!decodeFirstFrame) {
Class animatedImageClass = context[SDWebImageContextAnimatedImageClass];
// check whether we should use `SDAnimatedImage`
if ([animatedImageClass isSubclassOfClass:[UIImage class]] && [animatedImageClass conformsToProtocol:@protocol(SDAnimatedImage)]) {
image = [[animatedImageClass alloc] initWithData:imageData scale:scale options:coderOptions];
if (image) {
// Preload frames if supported
if (options & SDWebImagePreloadAllFrames && [image respondsToSelector:@selector(preloadAllFrames)]) {
[((id)image) preloadAllFrames];
}
} else {
// Check image class matching
if (options & SDWebImageMatchAnimatedImageClass) {
return nil;
}
}
}
}
if (!image) {
image = [[SDImageCodersManager sharedManager] decodedImageWithData:imageData options:coderOptions];
}
if (image) {
BOOL shouldDecode = !SD_OPTIONS_CONTAINS(options, SDWebImageAvoidDecodeImage);
if ([image.class conformsToProtocol:@protocol(SDAnimatedImage)]) {
// `SDAnimatedImage` do not decode
shouldDecode = NO;
} else if (image.sd_isAnimated) {
// animated image do not decode
shouldDecode = NO;
}
if (shouldDecode) {
BOOL shouldScaleDown = SD_OPTIONS_CONTAINS(options, SDWebImageScaleDownLargeImages);
if (shouldScaleDown) {
image = [SDImageCoderHelper decodedAndScaledDownImageWithImage:image limitBytes:0];
} else {
image = [SDImageCoderHelper decodedImageWithImage:image];
}
}
}
return image;
}
FAQ
-[SDWebImageDownloaderDecryptor decryptedDataWithData:response:]
内部调用私有block处理imageData,并返回NSData赋值给imageData。
-
Q1:
寻找 decryptor 初始化的位置,以及 block 的实现在哪 -
A1:
没有实现block的地方,默认情况下 decryptor 是nil,此为开发者来指定的属性
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
-
Q2:
SDImageLoader.h 和 SDImageLoader.m 文件的构成 -
A2:
SDImageLoader.h 只有两个函数的EXPORT声明,以及一个协议的声明;SDImageLoader.m 仅仅是两个函数的实现。
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
-
Q3:
SDImageCacheDefine 文件和 SDImageLoader 文件 -
A3:
解码代码基本一致,声明的协议不同。
解码:
SDImageLoader 的解码实现是SDImageLoaderDecodeImageData()
SDImageCacheDefine 的解码实现是SDImageCacheDecodeImageData()
协议:
SDImageLoader 声明的协议是@protocol SDImageLoader
;
SDImageCacheDefine 声明的协议是@protocol SDImageCache
;
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
-
Q4:
有多少种Coder? -
A4:
Coder大致分两类:苹果内建的和SD自己实现的;一共5种;
-
SDImageIOCoder
苹果内建,支持PNG, JPEG, TIFF解码,以及 GIF 第一帧的静态图。- 使用
UIImage *image = [[UIImage alloc] initWithData:data scale:scale];
进行解码
- 使用
SDImageIOAnimatedCoder
下面三种的基类SDImageGIFCoder
继承自SDImageIOAnimatedCoder
SDImageAPNGCoder
继承自SDImageIOAnimatedCoder
SDImageHEICCoder
继承自SDImageIOAnimatedCoder
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
-
Q5:
CGContextSetInterpolationQuality()
是在设置什么? -
A5:
在设置“插值质量”。对图片的尺寸进行缩放时,由于尺寸不同,所以在生成新图的过程中像素不可能是一一对应,因此会有插值操作。所谓插值即根据原图和目标图大小比例,结合原图像素信息生成的新的像素的过程。常见的插值算法有线性插值,双线性插值,立方卷积插值等。
一点点图片相关小知识
以下内容均来自苹果官方文档和官方示例,地址:Quartz 2D Programming Guide.
首先普及一下图片的关键小知识:
颜色空间(ColorSpace, CS)
颜色空间最直白的说就是平时所见到的例如 RGB、CMYK 这些个东西
颜色空间文档地址
像素格式(Pixel format)
由三个数值组成: Bits per pixel (bpp)、Bits per component (bpc)、Bytes per row
像素格式文档地址
颜色分量(Color Component);
颜色分量就是颜色空间的各个小组件。比如 RGB 颜色空间,R 是一个颜色分量,同理G和B都是颜色分量。
颜色空间对应的像素格式
每个颜色空间对应的像素格式是规定好的(定死的),即bits per pixel (bpp) 和 bits per component (bpc);下表就是各种颜色空间对应的像素格式:
ColorSpace | Pixel format | bitmap information constant | Availability |
---|---|---|---|
Null | 8 bpp, 8 bpc | kCGImageAlphaOnly | Mac OS X, iOS |
Gray | 8 bpp, 8 bpc | kCGImageAlphaNone | Mac OS X, iOS |
Gray | 8 bpp, 8 bpc | kCGImageAlphaOnly | Mac OS X, iOS |
RGB | 16 bpp, 5 bpc | kCGImageAlphaNoneSkipFirst | Mac OS X, iOS |
RGB | 32 bpp, 8 bpc | kCGImageAlphaNoneSkipFirst | Mac OS X, iOS |
RGB | 32 bpp, 8 bpc | kCGImageAlphaNoneSkipLast | Mac OS X, iOS |
RGB | 32 bpp, 8 bpc | kCGImageAlphaPremultipliedFirst | Mac OS X, iOS |
RGB | 32 bpp, 8 bpc | kCGImageAlphaPremultipliedLast | Mac OS X, iOS |
解释一下为什么上表中像素格式为啥缺了 Bytes per row 这个指标,是因为 Bytes per row 其实是通过计算确定的,Bytes per row = width * bytes per pixel
。 所以有了 bpp 也就同时有了 Bytes per row
参考解码大图的官方示例
解码大图的官方示例中定义了许多宏,可以让我们很快的熟悉图片的相关知识。哪些是可以定死的,哪些是可以通过计算获得的。
#define bytesPerMB 1048576.0f
#define bytesPerPixel 4.0f
#define pixelsPerMB ( bytesPerMB / bytesPerPixel ) // 262144 pixels, for 4 bytes per pixel.
#define destTotalPixels kDestImageSizeMB * pixelsPerMB
#define tileTotalPixels kSourceImageTileSizeMB * pixelsPerMB
#define destSeemOverlap 2.0f // the numbers of pixels to overlap the seems where tiles meet.
#define kDestImageSizeMB 60.0f // The resulting image will be (x)MB of uncompressed image data.
#define kSourceImageTileSizeMB 20.0f // The tile size will be (x)MB of uncompressed image data.
解码的核心函数 CGBitmapContextCreate()
翻译自 CGBitmapContextCreate() 函数的头文件注释,俺自己翻的,水平有限,请见谅
原文:The number of components for each pixel is specified by space
, which may also specify a destination color profile.
译文:每个像素的颜色分量数量由“space”指定,它也可以指定目标颜色配置文件。
原文:Note that the only legal case when space
can be NULL is when alpha is specified as kCGImageAlphaOnly.
译文:注意,space为NULL只有一种情况是合法的,就是当alpha被指定为kCGImageAlphaOnly。
原文:The number of bits for each component of a pixel is specified by bitsPerComponent
.
译文:一个颜色分量的比特数由“bitsPerComponent”指定。
原文:The number of bytes per pixel is equal to (bitsPerComponent * number of components + 7)/8
.
译文:一像素拥有的字节数 = (bitsPerComponent * number of components + 7)/8
原文:Each row of the bitmap consists of bytesPerRow
bytes, which must be at least width * bytes per pixel
bytes; in addition, bytesPerRow
must be an integer multiple of the number of bytes per pixel.
译文:位图每一行由bytesPerRow
字节构成,bytesPerRow
必须至少是width * bytes per pixel
字节。另外,bytesPerRow
必须是整型数字乘以一像素拥有的字节数。
原文:data
, if non-NULL, points to a block of memory at least bytesPerRow * height
bytes.
译文:data
如果非空,要指向一个至少是bytesPerRow * height
字节的内存块。
原文:If data
is NULL, the data for context is allocated automatically and freed when the context is deallocated.
译文:如果data
为空,context的data会被自动分配内存并且在context被销毁的时候释放。
原文:bitmapInfo
specifies whether the bitmap should contain an alpha channel and how it's to be generated, along with whether the components are floating-point or integer
译文:bitmapInfo
指定位图是否应该包含一个alpha通道,以及如何生成它,以及颜色分量是浮点数还是整数。