| #define DISPLAY(...) fprintf(stderr, __VA_ARGS__) |
| #define PRINT(...) fprintf(stdout, __VA_ARGS__) |
| #define DEBUG(l, ...) { if (g_displayLevel>=l) { DISPLAY(__VA_ARGS__); } } |
| #define FILE_CHUNK_SIZE 4 << 20 |
| #define MAX_NUM_JOBS 2 |
| #define stdinmark "/*stdin*\\" |
| #define stdoutmark "/*stdout*\\" |
| #define MAX_PATH 256 |
| #define DEFAULT_DISPLAY_LEVEL 1 |
| #define DEFAULT_COMPRESSION_LEVEL 6 |
| #define DEFAULT_ADAPT_PARAM 1 |
| typedef unsigned char BYTE; |
| |
| #include <stdio.h> /* fprintf */ |
| #include <stdlib.h> /* malloc, free */ |
| #include <pthread.h> /* pthread functions */ |
| #include <string.h> /* memset */ |
| #include "zstd_internal.h" |
| #include "util.h" |
| |
| static int g_displayLevel = DEFAULT_DISPLAY_LEVEL; |
| static unsigned g_compressionLevel = DEFAULT_COMPRESSION_LEVEL; |
| static unsigned g_displayStats = 0; |
| static UTIL_time_t g_startTime; |
| static size_t g_streamedSize = 0; |
| static unsigned g_useProgressBar = 0; |
| static UTIL_freq_t g_ticksPerSecond; |
| |
| typedef struct { |
| void* start; |
| size_t size; |
| } buffer_t; |
| |
| typedef struct { |
| size_t filled; |
| buffer_t buffer; |
| } inBuff_t; |
| |
| typedef struct { |
| unsigned waitCompressed; |
| unsigned waitReady; |
| unsigned waitWrite; |
| unsigned readyCounter; |
| unsigned compressedCounter; |
| unsigned writeCounter; |
| } cStat_t; |
| |
| typedef struct { |
| buffer_t src; |
| buffer_t dst; |
| buffer_t dict; |
| unsigned compressionLevel; |
| unsigned jobID; |
| unsigned lastJob; |
| size_t compressedSize; |
| } jobDescription; |
| |
| typedef struct { |
| unsigned compressionLevel; |
| unsigned numActiveThreads; |
| unsigned numJobs; |
| unsigned nextJobID; |
| unsigned threadError; |
| unsigned jobReadyID; |
| unsigned jobCompressedID; |
| unsigned jobWriteID; |
| unsigned allJobsCompleted; |
| unsigned adaptParam; |
| pthread_mutex_t jobCompressed_mutex; |
| pthread_cond_t jobCompressed_cond; |
| pthread_mutex_t jobReady_mutex; |
| pthread_cond_t jobReady_cond; |
| pthread_mutex_t allJobsCompleted_mutex; |
| pthread_cond_t allJobsCompleted_cond; |
| pthread_mutex_t jobWrite_mutex; |
| pthread_cond_t jobWrite_cond; |
| inBuff_t input; |
| cStat_t stats; |
| jobDescription* jobs; |
| FILE* dstFile; |
| ZSTD_CCtx* cctx; |
| } adaptCCtx; |
| |
| static void freeCompressionJobs(adaptCCtx* ctx) |
| { |
| unsigned u; |
| for (u=0; u<ctx->numJobs; u++) { |
| jobDescription job = ctx->jobs[u]; |
| free(job.dst.start); |
| free(job.src.start); |
| free(job.dict.start); |
| } |
| } |
| |
| static int freeCCtx(adaptCCtx* ctx) |
| { |
| if (!ctx) return 0; |
| int const compressedMutexError = pthread_mutex_destroy(&ctx->jobCompressed_mutex); |
| int const compressedCondError = pthread_cond_destroy(&ctx->jobCompressed_cond); |
| int const readyMutexError = pthread_mutex_destroy(&ctx->jobReady_mutex); |
| int const readyCondError = pthread_cond_destroy(&ctx->jobReady_cond); |
| int const allJobsMutexError = pthread_mutex_destroy(&ctx->allJobsCompleted_mutex); |
| int const allJobsCondError = pthread_cond_destroy(&ctx->allJobsCompleted_cond); |
| int const jobWriteMutexError = pthread_mutex_destroy(&ctx->jobWrite_mutex); |
| int const jobWriteCondError = pthread_cond_destroy(&ctx->jobWrite_cond); |
| int const fileCloseError = (ctx->dstFile != NULL && ctx->dstFile != stdout) ? fclose(ctx->dstFile) : 0; |
| int const cctxError = ZSTD_isError(ZSTD_freeCCtx(ctx->cctx)) ? 1 : 0; |
| free(ctx->input.buffer.start); |
| if (ctx->jobs){ |
| freeCompressionJobs(ctx); |
| free(ctx->jobs); |
| } |
| free(ctx); |
| return compressedMutexError | compressedCondError | readyMutexError | readyCondError | fileCloseError | allJobsMutexError | allJobsCondError | jobWriteMutexError | jobWriteCondError | cctxError; |
| } |
| |
| static adaptCCtx* createCCtx(unsigned numJobs, const char* const outFilename) |
| { |
| |
| adaptCCtx* ctx = malloc(sizeof(adaptCCtx)); |
| if (ctx == NULL) { |
| DISPLAY("Error: could not allocate space for context\n"); |
| return NULL; |
| } |
| memset(ctx, 0, sizeof(adaptCCtx)); |
| ctx->compressionLevel = g_compressionLevel; |
| pthread_mutex_init(&ctx->jobCompressed_mutex, NULL); |
| pthread_cond_init(&ctx->jobCompressed_cond, NULL); |
| pthread_mutex_init(&ctx->jobReady_mutex, NULL); |
| pthread_cond_init(&ctx->jobReady_cond, NULL); |
| pthread_mutex_init(&ctx->allJobsCompleted_mutex, NULL); |
| pthread_cond_init(&ctx->allJobsCompleted_cond, NULL); |
| pthread_mutex_init(&ctx->jobWrite_mutex, NULL); |
| pthread_cond_init(&ctx->jobWrite_cond, NULL); |
| ctx->numJobs = numJobs; |
| ctx->jobReadyID = 0; |
| ctx->jobCompressedID = 0; |
| ctx->jobWriteID = 0; |
| ctx->jobs = calloc(1, numJobs*sizeof(jobDescription)); |
| /* initializing jobs */ |
| { |
| unsigned jobNum; |
| for (jobNum=0; jobNum<numJobs; jobNum++) { |
| jobDescription* job = &ctx->jobs[jobNum]; |
| job->src.start = malloc(FILE_CHUNK_SIZE); |
| job->dst.start = malloc(ZSTD_compressBound(FILE_CHUNK_SIZE)); |
| job->dict.start = malloc(FILE_CHUNK_SIZE); |
| job->lastJob = 0; |
| if (!job->src.start || !job->dst.start || !job->dict.start) { |
| DISPLAY("Could not allocate buffers for jobs\n"); |
| freeCCtx(ctx); |
| return NULL; |
| } |
| } |
| } |
| ctx->nextJobID = 0; |
| ctx->threadError = 0; |
| ctx->allJobsCompleted = 0; |
| ctx->adaptParam = DEFAULT_ADAPT_PARAM; |
| ctx->cctx = ZSTD_createCCtx(); |
| ctx->input.filled = 0; |
| ctx->input.buffer.size = 2 * FILE_CHUNK_SIZE; |
| ctx->input.buffer.start = malloc(ctx->input.buffer.size); |
| if (!ctx->input.buffer.start) { |
| DISPLAY("Error: could not allocate input buffer\n"); |
| freeCCtx(ctx); |
| return NULL; |
| } |
| if (!ctx->cctx) { |
| DISPLAY("Error: could not allocate ZSTD_CCtx\n"); |
| freeCCtx(ctx); |
| return NULL; |
| } |
| if (!ctx->jobs) { |
| DISPLAY("Error: could not allocate space for jobs during context creation\n"); |
| freeCCtx(ctx); |
| return NULL; |
| } |
| { |
| unsigned const stdoutUsed = !strcmp(outFilename, stdoutmark); |
| FILE* dstFile = stdoutUsed ? stdout : fopen(outFilename, "wb"); |
| if (dstFile == NULL) { |
| DISPLAY("Error: could not open output file\n"); |
| freeCCtx(ctx); |
| return NULL; |
| } |
| ctx->dstFile = dstFile; |
| } |
| return ctx; |
| } |
| |
| |
| |
| static void waitUntilAllJobsCompleted(adaptCCtx* ctx) |
| { |
| pthread_mutex_lock(&ctx->allJobsCompleted_mutex); |
| while (ctx->allJobsCompleted == 0) { |
| pthread_cond_wait(&ctx->allJobsCompleted_cond, &ctx->allJobsCompleted_mutex); |
| } |
| pthread_mutex_unlock(&ctx->allJobsCompleted_mutex); |
| } |
| |
| static unsigned adaptCompressionLevel(adaptCCtx* ctx) |
| { |
| unsigned reset = 0; |
| unsigned const allSlow = ctx->adaptParam < ctx->stats.compressedCounter && ctx->adaptParam < ctx->stats.writeCounter && ctx->adaptParam < ctx->stats.readyCounter ? 1 : 0; |
| unsigned const compressWaiting = ctx->adaptParam < ctx->stats.readyCounter ? 1 : 0; |
| unsigned const writeWaiting = ctx->adaptParam < ctx->stats.compressedCounter ? 1 : 0; |
| unsigned const createWaiting = ctx->adaptParam < ctx->stats.writeCounter ? 1 : 0; |
| unsigned const writeSlow = ((compressWaiting && createWaiting) || (createWaiting && !writeWaiting)) ? 1 : 0; |
| unsigned const compressSlow = ((writeWaiting && createWaiting) || (writeWaiting && !compressWaiting)) ? 1 : 0; |
| unsigned const createSlow = ((compressWaiting && writeWaiting) || (compressWaiting && !createWaiting)) ? 1 : 0; |
| DEBUG(3, "ready: %u compressed: %u write: %u\n", ctx->stats.readyCounter, ctx->stats.compressedCounter, ctx->stats.writeCounter); |
| if (allSlow) { |
| reset = 1; |
| } |
| else if ((writeSlow || createSlow) && ctx->compressionLevel < (unsigned)ZSTD_maxCLevel()) { |
| DEBUG(3, "increasing compression level %u\n", ctx->compressionLevel); |
| ctx->compressionLevel++; |
| reset = 1; |
| } |
| else if (compressSlow && ctx->compressionLevel > 1) { |
| DEBUG(3, "decreasing compression level %u\n", ctx->compressionLevel); |
| ctx->compressionLevel--; |
| reset = 1; |
| } |
| if (reset) { |
| ctx->stats.readyCounter = 0; |
| ctx->stats.writeCounter = 0; |
| ctx->stats.compressedCounter = 0; |
| } |
| return ctx->compressionLevel; |
| } |
| |
| static void* compressionThread(void* arg) |
| { |
| adaptCCtx* ctx = (adaptCCtx*)arg; |
| unsigned currJob = 0; |
| for ( ; ; ) { |
| unsigned const currJobIndex = currJob % ctx->numJobs; |
| jobDescription* job = &ctx->jobs[currJobIndex]; |
| DEBUG(3, "compressionThread(): waiting on job ready\n"); |
| pthread_mutex_lock(&ctx->jobReady_mutex); |
| while(currJob + 1 > ctx->jobReadyID) { |
| ctx->stats.waitReady++; |
| ctx->stats.readyCounter++; |
| DEBUG(3, "waiting on job ready, nextJob: %u\n", currJob); |
| pthread_cond_wait(&ctx->jobReady_cond, &ctx->jobReady_mutex); |
| } |
| pthread_mutex_unlock(&ctx->jobReady_mutex); |
| DEBUG(3, "compressionThread(): continuing after job ready\n"); |
| /* compress the data */ |
| { |
| unsigned const cLevel = adaptCompressionLevel(ctx); |
| DEBUG(3, "cLevel used: %u\n", cLevel); |
| DEBUG(2, "dictSize: %zu, srcSize: %zu\n", job->dict.size, job->src.size); |
| /* begin compression */ |
| { |
| size_t const dictModeError = ZSTD_setCCtxParameter(ctx->cctx, ZSTD_p_forceRawDict, 1); |
| size_t const initError = ZSTD_compressBegin_usingDict(ctx->cctx, job->dict.start, job->dict.size, cLevel); |
| size_t const windowSizeError = ZSTD_setCCtxParameter(ctx->cctx, ZSTD_p_forceWindow, 1); |
| if (ZSTD_isError(dictModeError) || ZSTD_isError(initError) || ZSTD_isError(windowSizeError)) { |
| DISPLAY("Error: something went wrong while starting compression\n"); |
| ctx->threadError = 1; |
| return arg; |
| } |
| } |
| |
| /* continue compression */ |
| if (currJob != 0) { /* not first job */ |
| size_t const hSize = ZSTD_compressContinue(ctx->cctx, job->dst.start, job->dst.size, job->src.start, job->src.size); |
| if (ZSTD_isError(hSize)) { |
| DISPLAY("Error: something went wrong while continuing compression\n"); |
| job->compressedSize = hSize; |
| ctx->threadError = 1; |
| return arg; |
| } |
| ZSTD_invalidateRepCodes(ctx->cctx); |
| } |
| job->compressedSize = (job->lastJob) ? |
| ZSTD_compressEnd (ctx->cctx, job->dst.start, job->dst.size, job->src.start, job->src.size) : |
| ZSTD_compressContinue(ctx->cctx, job->dst.start, job->dst.size, job->src.start, job->src.size); |
| if (ZSTD_isError(job->compressedSize)) { |
| DISPLAY("Error: something went wrong during compression: %s\n", ZSTD_getErrorName(job->compressedSize)); |
| ctx->threadError = 1; |
| return arg; |
| } |
| } |
| pthread_mutex_lock(&ctx->jobCompressed_mutex); |
| ctx->jobCompressedID++; |
| DEBUG(3, "signaling for job %u\n", currJob); |
| pthread_cond_signal(&ctx->jobCompressed_cond); |
| pthread_mutex_unlock(&ctx->jobCompressed_mutex); |
| DEBUG(3, "finished job compression %u\n", currJob); |
| currJob++; |
| if (job->lastJob || ctx->threadError) { |
| /* finished compressing all jobs */ |
| DEBUG(3, "all jobs finished compressing\n"); |
| break; |
| } |
| } |
| return arg; |
| } |
| |
| static void displayProgress(unsigned jobDoneID, unsigned cLevel, unsigned last) |
| { |
| if (!g_useProgressBar) return; |
| UTIL_time_t currTime; |
| UTIL_getTime(&currTime); |
| double const timeElapsed = (double)(UTIL_getSpanTimeMicro(g_ticksPerSecond, g_startTime, currTime) / 1000.0); |
| double const sizeMB = (double)g_streamedSize / (1 << 20); |
| double const avgCompRate = sizeMB * 1000 / timeElapsed; |
| fprintf(stdout, "\r| %4u jobs completed | Current Compresion Level: %2u | Time Elapsed: %5.0f ms | Data Size: %7.1f MB | Avg Compression Rate: %6.2f MB/s |", jobDoneID, cLevel, timeElapsed, sizeMB, avgCompRate); |
| if (last) { |
| fprintf(stdout, "\n"); |
| } |
| else { |
| fflush(stdout); |
| } |
| } |
| |
| static void* outputThread(void* arg) |
| { |
| adaptCCtx* ctx = (adaptCCtx*)arg; |
| |
| unsigned currJob = 0; |
| for ( ; ; ) { |
| unsigned const currJobIndex = currJob % ctx->numJobs; |
| jobDescription* job = &ctx->jobs[currJobIndex]; |
| DEBUG(3, "outputThread(): waiting on job compressed\n"); |
| pthread_mutex_lock(&ctx->jobCompressed_mutex); |
| while (currJob + 1 > ctx->jobCompressedID) { |
| ctx->stats.waitCompressed++; |
| ctx->stats.compressedCounter++; |
| DEBUG(3, "waiting on job compressed, nextJob: %u\n", currJob); |
| pthread_cond_wait(&ctx->jobCompressed_cond, &ctx->jobCompressed_mutex); |
| } |
| pthread_mutex_unlock(&ctx->jobCompressed_mutex); |
| DEBUG(3, "outputThread(): continuing after job compressed\n"); |
| { |
| size_t const compressedSize = job->compressedSize; |
| if (ZSTD_isError(compressedSize)) { |
| DISPLAY("Error: an error occurred during compression\n"); |
| ctx->threadError = 1; |
| return arg; |
| } |
| { |
| size_t const writeSize = fwrite(job->dst.start, 1, compressedSize, ctx->dstFile); |
| if (writeSize != compressedSize) { |
| DISPLAY("Error: an error occurred during file write operation\n"); |
| ctx->threadError = 1; |
| return arg; |
| } |
| } |
| } |
| DEBUG(3, "finished job write %u\n", currJob); |
| currJob++; |
| displayProgress(currJob, ctx->compressionLevel, job->lastJob); |
| DEBUG(3, "locking job write mutex\n"); |
| pthread_mutex_lock(&ctx->jobWrite_mutex); |
| ctx->jobWriteID++; |
| pthread_cond_signal(&ctx->jobWrite_cond); |
| pthread_mutex_unlock(&ctx->jobWrite_mutex); |
| DEBUG(3, "unlocking job write mutex\n"); |
| |
| if (job->lastJob || ctx->threadError) { |
| /* finished with all jobs */ |
| DEBUG(3, "all jobs finished writing\n"); |
| pthread_mutex_lock(&ctx->allJobsCompleted_mutex); |
| ctx->allJobsCompleted = 1; |
| pthread_cond_signal(&ctx->allJobsCompleted_cond); |
| pthread_mutex_unlock(&ctx->allJobsCompleted_mutex); |
| break; |
| } |
| } |
| return arg; |
| } |
| |
| static int createCompressionJob(adaptCCtx* ctx, size_t srcSize, int last) |
| { |
| unsigned const nextJob = ctx->nextJobID; |
| unsigned const nextJobIndex = nextJob % ctx->numJobs; |
| jobDescription* job = &ctx->jobs[nextJobIndex]; |
| DEBUG(3, "createCompressionJob(): wait for job write\n"); |
| pthread_mutex_lock(&ctx->jobWrite_mutex); |
| DEBUG(3, "Creating new compression job -- nextJob: %u, jobCompressedID: %u, jobWriteID: %u, numJObs: %u\n", nextJob,ctx->jobCompressedID, ctx->jobWriteID, ctx->numJobs); |
| while (nextJob - ctx->jobWriteID >= ctx->numJobs) { |
| ctx->stats.waitWrite++; |
| ctx->stats.writeCounter++; |
| DEBUG(3, "waiting on job Write, nextJob: %u\n", nextJob); |
| pthread_cond_wait(&ctx->jobWrite_cond, &ctx->jobWrite_mutex); |
| } |
| pthread_mutex_unlock(&ctx->jobWrite_mutex); |
| DEBUG(3, "createCompressionJob(): continuing after job write\n"); |
| |
| |
| job->compressionLevel = ctx->compressionLevel; |
| job->src.size = srcSize; |
| job->dst.size = ZSTD_compressBound(srcSize); |
| job->jobID = nextJob; |
| job->lastJob = last; |
| memcpy(job->src.start, ctx->input.buffer.start + ctx->input.filled, srcSize); |
| job->dict.size = ctx->input.filled; |
| memcpy(job->dict.start, ctx->input.buffer.start, ctx->input.filled); |
| pthread_mutex_lock(&ctx->jobReady_mutex); |
| ctx->jobReadyID++; |
| pthread_cond_signal(&ctx->jobReady_cond); |
| pthread_mutex_unlock(&ctx->jobReady_mutex); |
| DEBUG(3, "finished job creation %u\n", nextJob); |
| ctx->nextJobID++; |
| DEBUG(3, "filled: %zu, srcSize: %zu\n", ctx->input.filled, srcSize); |
| /* if not on the last job, reuse data as dictionary in next job */ |
| if (!last) { |
| size_t const newDictSize = srcSize/16; |
| size_t const oldDictSize = ctx->input.filled; |
| memmove(ctx->input.buffer.start, ctx->input.buffer.start + oldDictSize + srcSize - newDictSize, newDictSize); |
| ctx->input.filled = newDictSize; |
| } |
| return 0; |
| } |
| |
| static void printStats(cStat_t stats) |
| { |
| DISPLAY("========STATISTICS========\n"); |
| DISPLAY("# times waited on job ready: %u\n", stats.waitReady); |
| DISPLAY("# times waited on job compressed: %u\n", stats.waitCompressed); |
| DISPLAY("# times waited on job Write: %u\n\n", stats.waitWrite); |
| } |
| |
| static int compressFilename(const char* const srcFilename, const char* const dstFilenameOrNull) |
| { |
| unsigned const stdinUsed = !strcmp(srcFilename, stdinmark); |
| FILE* const srcFile = stdinUsed ? stdin : fopen(srcFilename, "rb"); |
| const char* const outFilenameIntermediate = (stdinUsed && !dstFilenameOrNull) ? stdoutmark : dstFilenameOrNull; |
| const char* outFilename = outFilenameIntermediate; |
| char fileAndSuffix[MAX_PATH]; |
| size_t const numJobs = MAX_NUM_JOBS; |
| int ret = 0; |
| adaptCCtx* ctx = NULL; |
| UTIL_getTime(&g_startTime); |
| g_streamedSize = 0; |
| |
| if (!outFilenameIntermediate) { |
| if (snprintf(fileAndSuffix, MAX_PATH, "%s.zst", srcFilename) + 1 > MAX_PATH) { |
| DISPLAY("Error: output filename is too long\n"); |
| ret = 1; |
| goto cleanup; |
| } |
| outFilename = fileAndSuffix; |
| } |
| |
| /* checking for errors */ |
| if (!srcFilename || !outFilename || !srcFile) { |
| DISPLAY("Error: initial variables could not be allocated\n"); |
| ret = 1; |
| goto cleanup; |
| } |
| |
| /* creating context */ |
| ctx = createCCtx(numJobs, outFilename); |
| if (ctx == NULL) { |
| ret = 1; |
| goto cleanup; |
| } |
| |
| /* create output thread */ |
| { |
| pthread_t out; |
| if (pthread_create(&out, NULL, &outputThread, ctx)) { |
| DISPLAY("Error: could not create output thread\n"); |
| ret = 1; |
| goto cleanup; |
| } |
| } |
| |
| /* create compression thread */ |
| { |
| pthread_t compression; |
| if (pthread_create(&compression, NULL, &compressionThread, ctx)) { |
| DISPLAY("Error: could not create compression thread\n"); |
| ret = 1; |
| goto cleanup; |
| } |
| } |
| |
| /* creating jobs */ |
| for ( ; ; ) { |
| size_t const readSize = fread(ctx->input.buffer.start + ctx->input.filled, 1, FILE_CHUNK_SIZE, srcFile); |
| if (readSize != FILE_CHUNK_SIZE && !feof(srcFile)) { |
| DISPLAY("Error: problem occurred during read from src file\n"); |
| ctx->threadError = 1; |
| ret = 1; |
| goto cleanup; |
| } |
| g_streamedSize += readSize; |
| /* reading was fine, now create the compression job */ |
| { |
| int const last = feof(srcFile); |
| int const error = createCompressionJob(ctx, readSize, last); |
| if (error != 0) { |
| ret = error; |
| ctx->threadError = 1; |
| goto cleanup; |
| } |
| } |
| if (feof(srcFile)) { |
| DEBUG(3, "THE STREAM OF DATA ENDED %u\n", ctx->nextJobID); |
| break; |
| } |
| } |
| |
| cleanup: |
| waitUntilAllJobsCompleted(ctx); |
| if (g_displayStats) printStats(ctx->stats); |
| /* file compression completed */ |
| ret |= (srcFile != NULL) ? fclose(srcFile) : 0; |
| ret |= (ctx != NULL) ? freeCCtx(ctx) : 0; |
| return ret; |
| } |
| |
| static int compressFilenames(const char** filenameTable, unsigned numFiles, unsigned forceStdout) |
| { |
| int ret = 0; |
| unsigned fileNum; |
| for (fileNum=0; fileNum<numFiles; fileNum++) { |
| const char* filename = filenameTable[fileNum]; |
| if (!forceStdout) { |
| ret |= compressFilename(filename, NULL); |
| } |
| else { |
| ret |= compressFilename(filename, stdoutmark); |
| } |
| |
| } |
| return ret; |
| } |
| |
| /*! readU32FromChar() : |
| @return : unsigned integer value read from input in `char` format |
| allows and interprets K, KB, KiB, M, MB and MiB suffix. |
| Will also modify `*stringPtr`, advancing it to position where it stopped reading. |
| Note : function result can overflow if digit string > MAX_UINT */ |
| static unsigned readU32FromChar(const char** stringPtr) |
| { |
| unsigned result = 0; |
| while ((**stringPtr >='0') && (**stringPtr <='9')) |
| result *= 10, result += **stringPtr - '0', (*stringPtr)++ ; |
| if ((**stringPtr=='K') || (**stringPtr=='M')) { |
| result <<= 10; |
| if (**stringPtr=='M') result <<= 10; |
| (*stringPtr)++ ; |
| if (**stringPtr=='i') (*stringPtr)++; |
| if (**stringPtr=='B') (*stringPtr)++; |
| } |
| return result; |
| } |
| |
| static void help() |
| { |
| PRINT("Usage:\n"); |
| PRINT(" ./multi [options] [file(s)]\n"); |
| PRINT("\n"); |
| PRINT("Options:\n"); |
| PRINT(" -oFILE : specify the output file name\n"); |
| PRINT(" -v : display debug information\n"); |
| PRINT(" -i# : provide initial compression level\n"); |
| PRINT(" -s : display information stats\n"); |
| PRINT(" -h : display help/information\n"); |
| } |
| /* return 0 if successful, else return error */ |
| int main(int argCount, const char* argv[]) |
| { |
| const char* outFilename = NULL; |
| const char** filenameTable = (const char**)malloc(argCount*sizeof(const char*)); |
| unsigned filenameIdx = 0; |
| filenameTable[0] = stdinmark; |
| unsigned forceStdout = 0; |
| int ret = 0; |
| int argNum; |
| |
| UTIL_initTimer(&g_ticksPerSecond); |
| |
| if (filenameTable == NULL) { |
| DISPLAY("Error: could not allocate sapce for filename table.\n"); |
| return 1; |
| } |
| |
| for (argNum=1; argNum<argCount; argNum++) { |
| const char* argument = argv[argNum]; |
| |
| /* output filename designated with "-o" */ |
| if (argument[0]=='-' && strlen(argument) > 1) { |
| switch (argument[1]) { |
| case 'o': |
| argument += 2; |
| outFilename = argument; |
| break; |
| case 'v': |
| g_displayLevel++; |
| break; |
| case 'i': |
| argument += 2; |
| g_compressionLevel = readU32FromChar(&argument); |
| DEBUG(3, "g_compressionLevel: %u\n", g_compressionLevel); |
| break; |
| case 's': |
| g_displayStats = 1; |
| break; |
| case 'h': |
| help(); |
| goto _main_exit; |
| case 'p': |
| g_useProgressBar = 1; |
| break; |
| case 'c': |
| forceStdout = 1; |
| outFilename = stdoutmark; |
| break; |
| default: |
| DISPLAY("Error: invalid argument provided\n"); |
| ret = 1; |
| goto _main_exit; |
| } |
| continue; |
| } |
| |
| /* regular files to be compressed */ |
| filenameTable[filenameIdx++] = argument; |
| } |
| |
| /* error checking with number of files */ |
| if (filenameIdx > 1 && (outFilename != NULL && strcmp(outFilename, stdoutmark))) { |
| DISPLAY("Error: multiple input files provided, cannot use specified output file\n"); |
| ret = 1; |
| goto _main_exit; |
| } |
| |
| /* compress files */ |
| if (filenameIdx <= 1) { |
| ret |= compressFilename(filenameTable[0], outFilename); |
| } |
| else { |
| ret |= compressFilenames(filenameTable, filenameIdx, forceStdout); |
| } |
| _main_exit: |
| free(filenameTable); |
| return ret; |
| } |