diff options
author | Stefan Radomski <radomski@tk.informatik.tu-darmstadt.de> | 2013-10-16 18:35:17 (GMT) |
---|---|---|
committer | Stefan Radomski <radomski@tk.informatik.tu-darmstadt.de> | 2013-10-16 18:35:17 (GMT) |
commit | 60c543c1d8c6699cc96829f15b16ecc60ac0b410 (patch) | |
tree | f0087f620d45246acf0268da539cb815796d88c0 /test | |
parent | 4d186e24e44fe07f04708c2dc38387534c02c7fd (diff) | |
download | uscxml-60c543c1d8c6699cc96829f15b16ecc60ac0b410.zip uscxml-60c543c1d8c6699cc96829f15b16ecc60ac0b410.tar.gz uscxml-60c543c1d8c6699cc96829f15b16ecc60ac0b410.tar.bz2 |
Added license information and bug fixes
Diffstat (limited to 'test')
-rw-r--r-- | test/CMakeLists.txt | 6 | ||||
-rw-r--r-- | test/src/test-cmdline-parsing.cpp | 8 | ||||
-rw-r--r-- | test/src/test-ffmpeg.cpp | 518 | ||||
-rw-r--r-- | test/src/test-instant-messaging.cpp | 52 | ||||
-rw-r--r-- | test/src/test-url.cpp | 2 | ||||
-rw-r--r-- | test/src/test-w3c.cpp | 2 |
6 files changed, 281 insertions, 307 deletions
diff --git a/test/CMakeLists.txt b/test/CMakeLists.txt index 261817b..eff801c 100644 --- a/test/CMakeLists.txt +++ b/test/CMakeLists.txt @@ -118,9 +118,9 @@ set_target_properties(scxml-test-framework-client PROPERTIES FOLDER "Tests") # add_test(test-curl-multi-api ${CMAKE_RUNTIME_OUTPUT_DIRECTORY}/test-completion) # set_target_properties(test-curl-multi-api PROPERTIES FOLDER "Tests") - add_executable(test-w3c src/test-w3c.cpp) - target_link_libraries(test-w3c uscxml) - set_target_properties(test-w3c PROPERTIES FOLDER "Tests") +add_executable(test-w3c src/test-w3c.cpp) +target_link_libraries(test-w3c uscxml) +set_target_properties(test-w3c PROPERTIES FOLDER "Tests") file(GLOB_RECURSE W3C_TESTS samples/w3c/*.scxml diff --git a/test/src/test-cmdline-parsing.cpp b/test/src/test-cmdline-parsing.cpp index 9ff094a..a610248 100644 --- a/test/src/test-cmdline-parsing.cpp +++ b/test/src/test-cmdline-parsing.cpp @@ -33,8 +33,6 @@ int main(int argc, char** argv) { assert(options.logLevel == 10); assert(!options.withHTTP); assert(!options); // invalid as no SCXML document is given - - optind = 0; } if (true) { @@ -50,8 +48,6 @@ int main(int argc, char** argv) { assert(options.verbose); assert(options.interpreters.size() == 1); assert(options.interpreters.find("/foo/bar.scxml") != options.interpreters.end()); - - optind = 0; } if (true) { @@ -76,8 +72,6 @@ int main(int argc, char** argv) { assert(!options.interpreters["/foo/bar1.scxml"]->withHTTP); assert(options.interpreters["/foo/bar2.scxml"]->withHTTP); assert(!options.interpreters["/foo/bar3.scxml"]->withHTTP); - - optind = 0; } if (true) { @@ -97,8 +91,6 @@ int main(int argc, char** argv) { assert(options.interpreters.find("/foo/bar1.scxml") != options.interpreters.end()); assert(options.interpreters["/foo/bar1.scxml"]->additionalParameters.find("vrml-path") != options.interpreters["/foo/bar1.scxml"]->additionalParameters.end()); assert(options.interpreters["/foo/bar1.scxml"]->additionalParameters.find("tmp-path") != options.interpreters["/foo/bar1.scxml"]->additionalParameters.end()); - - optind = 0; } return EXIT_SUCCESS; diff --git a/test/src/test-ffmpeg.cpp b/test/src/test-ffmpeg.cpp index c9dcfe1..a8660e5 100644 --- a/test/src/test-ffmpeg.cpp +++ b/test/src/test-ffmpeg.cpp @@ -53,71 +53,70 @@ static int sws_flags = SWS_BICUBIC; /* Add an output stream. */ static AVStream *add_stream(AVFormatContext *oc, AVCodec **codec, - enum AVCodecID codec_id) -{ - AVCodecContext *c; - AVStream *st; - - /* find the encoder */ - *codec = avcodec_find_encoder(codec_id); - if (!(*codec)) { - fprintf(stderr, "Could not find encoder for '%s'\n", - avcodec_get_name(codec_id)); - exit(1); - } - - st = avformat_new_stream(oc, *codec); - if (!st) { - fprintf(stderr, "Could not allocate stream\n"); - exit(1); - } - st->id = oc->nb_streams-1; - c = st->codec; - - switch ((*codec)->type) { - case AVMEDIA_TYPE_AUDIO: - c->sample_fmt = AV_SAMPLE_FMT_FLTP; - c->bit_rate = 64000; - c->sample_rate = 44100; - c->channels = 2; - break; - - case AVMEDIA_TYPE_VIDEO: - c->codec_id = codec_id; - - c->bit_rate = 400000; - /* Resolution must be a multiple of two. */ - c->width = 352; - c->height = 288; - /* timebase: This is the fundamental unit of time (in seconds) in terms - * of which frame timestamps are represented. For fixed-fps content, - * timebase should be 1/framerate and timestamp increments should be - * identical to 1. */ - c->time_base.den = STREAM_FRAME_RATE; - c->time_base.num = 1; - c->gop_size = 12; /* emit one intra frame every twelve frames at most */ - c->pix_fmt = STREAM_PIX_FMT; - if (c->codec_id == AV_CODEC_ID_MPEG2VIDEO) { - /* just for testing, we also add B frames */ - c->max_b_frames = 2; - } - if (c->codec_id == AV_CODEC_ID_MPEG1VIDEO) { - /* Needed to avoid using macroblocks in which some coeffs overflow. - * This does not happen with normal video, it just happens here as - * the motion of the chroma plane does not match the luma plane. */ - c->mb_decision = 2; - } - break; - - default: - break; - } - - /* Some formats want stream headers to be separate. */ - if (oc->oformat->flags & AVFMT_GLOBALHEADER) - c->flags |= CODEC_FLAG_GLOBAL_HEADER; - - return st; + enum AVCodecID codec_id) { + AVCodecContext *c; + AVStream *st; + + /* find the encoder */ + *codec = avcodec_find_encoder(codec_id); + if (!(*codec)) { + fprintf(stderr, "Could not find encoder for '%s'\n", + avcodec_get_name(codec_id)); + exit(1); + } + + st = avformat_new_stream(oc, *codec); + if (!st) { + fprintf(stderr, "Could not allocate stream\n"); + exit(1); + } + st->id = oc->nb_streams-1; + c = st->codec; + + switch ((*codec)->type) { + case AVMEDIA_TYPE_AUDIO: + c->sample_fmt = AV_SAMPLE_FMT_FLTP; + c->bit_rate = 64000; + c->sample_rate = 44100; + c->channels = 2; + break; + + case AVMEDIA_TYPE_VIDEO: + c->codec_id = codec_id; + + c->bit_rate = 400000; + /* Resolution must be a multiple of two. */ + c->width = 352; + c->height = 288; + /* timebase: This is the fundamental unit of time (in seconds) in terms + * of which frame timestamps are represented. For fixed-fps content, + * timebase should be 1/framerate and timestamp increments should be + * identical to 1. */ + c->time_base.den = STREAM_FRAME_RATE; + c->time_base.num = 1; + c->gop_size = 12; /* emit one intra frame every twelve frames at most */ + c->pix_fmt = STREAM_PIX_FMT; + if (c->codec_id == AV_CODEC_ID_MPEG2VIDEO) { + /* just for testing, we also add B frames */ + c->max_b_frames = 2; + } + if (c->codec_id == AV_CODEC_ID_MPEG1VIDEO) { + /* Needed to avoid using macroblocks in which some coeffs overflow. + * This does not happen with normal video, it just happens here as + * the motion of the chroma plane does not match the luma plane. */ + c->mb_decision = 2; + } + break; + + default: + break; + } + + /* Some formats want stream headers to be separate. */ + if (oc->oformat->flags & AVFMT_GLOBALHEADER) + c->flags |= CODEC_FLAG_GLOBAL_HEADER; + + return st; } @@ -128,60 +127,58 @@ static AVFrame *frame; static AVPicture src_picture, dst_picture; static int frame_count; -static void open_video(AVFormatContext *oc, AVCodec *codec, AVStream *st) -{ - int ret; - AVCodecContext *c = st->codec; - - /* open the codec */ - ret = avcodec_open2(c, codec, NULL); - if (ret < 0) { - fprintf(stderr, "Could not open video codec: %s\n", av_err2str(ret)); - exit(1); - } - - /* allocate and init a re-usable frame */ - frame = avcodec_alloc_frame(); - if (!frame) { - fprintf(stderr, "Could not allocate video frame\n"); - exit(1); - } - - /* Allocate the encoded raw picture. */ - ret = avpicture_alloc(&dst_picture, c->pix_fmt, c->width, c->height); - if (ret < 0) { - fprintf(stderr, "Could not allocate picture: %s\n", av_err2str(ret)); - exit(1); - } - - /* If the output format is not YUV420P, then a temporary YUV420P - * picture is needed too. It is then converted to the required - * output format. */ - if (c->pix_fmt != BMP_FORMAT) { - ret = avpicture_alloc(&src_picture, BMP_FORMAT, c->width, c->height); - if (ret < 0) { - fprintf(stderr, "Could not allocate temporary picture: %s\n", - av_err2str(ret)); - exit(1); - } - } - - /* copy data and linesize picture pointers to frame */ - *((AVPicture *)frame) = dst_picture; +static void open_video(AVFormatContext *oc, AVCodec *codec, AVStream *st) { + int ret; + AVCodecContext *c = st->codec; + + /* open the codec */ + ret = avcodec_open2(c, codec, NULL); + if (ret < 0) { + fprintf(stderr, "Could not open video codec: %s\n", av_err2str(ret)); + exit(1); + } + + /* allocate and init a re-usable frame */ + frame = avcodec_alloc_frame(); + if (!frame) { + fprintf(stderr, "Could not allocate video frame\n"); + exit(1); + } + + /* Allocate the encoded raw picture. */ + ret = avpicture_alloc(&dst_picture, c->pix_fmt, c->width, c->height); + if (ret < 0) { + fprintf(stderr, "Could not allocate picture: %s\n", av_err2str(ret)); + exit(1); + } + + /* If the output format is not YUV420P, then a temporary YUV420P + * picture is needed too. It is then converted to the required + * output format. */ + if (c->pix_fmt != BMP_FORMAT) { + ret = avpicture_alloc(&src_picture, BMP_FORMAT, c->width, c->height); + if (ret < 0) { + fprintf(stderr, "Could not allocate temporary picture: %s\n", + av_err2str(ret)); + exit(1); + } + } + + /* copy data and linesize picture pointers to frame */ + *((AVPicture *)frame) = dst_picture; } /* Prepare a dummy image. */ static void fill_rgba_image(AVPicture *pict, int frame_index, - int width, int height) -{ + int width, int height) { int x, y, i; - + i = frame_index; - + std::stringstream ssFilename; ssFilename << "/Users/sradomski/Desktop/ctrl/" << (i % 125) << ".bmp"; - + std::ifstream file(ssFilename.str().c_str()); file.seekg(0, std::ios::end); @@ -190,7 +187,7 @@ static void fill_rgba_image(AVPicture *pict, int frame_index, char* buffer = (char*)malloc(length); file.read(buffer, length); - + uint32_t offset = 0; offset += buffer[10] << 0; offset += buffer[11] << 8; @@ -205,167 +202,164 @@ static void fill_rgba_image(AVPicture *pict, int frame_index, pict->data[0][y * pict->linesize[0] + x * 3 + 2] = buffer[offset++]; } } - + free(buffer); } -static void write_video_frame(AVFormatContext *oc, AVStream *st) -{ - int ret; - static struct SwsContext *sws_ctx; - AVCodecContext *c = st->codec; - - if (c->pix_fmt != BMP_FORMAT) { - /* as we only generate a YUV420P picture, we must convert it - * to the codec pixel format if needed */ - if (!sws_ctx) { - sws_ctx = sws_getContext(c->width, c->height, BMP_FORMAT, - c->width, c->height, c->pix_fmt, - sws_flags, NULL, NULL, NULL); - if (!sws_ctx) { - fprintf(stderr, - "Could not initialize the conversion context\n"); - exit(1); - } - } - fill_rgba_image(&src_picture, frame_count, c->width, c->height); - sws_scale(sws_ctx, - (const uint8_t * const *)src_picture.data, src_picture.linesize, - 0, c->height, dst_picture.data, dst_picture.linesize); - } else { - fill_rgba_image(&dst_picture, frame_count, c->width, c->height); +static void write_video_frame(AVFormatContext *oc, AVStream *st) { + int ret; + static struct SwsContext *sws_ctx; + AVCodecContext *c = st->codec; + + if (c->pix_fmt != BMP_FORMAT) { + /* as we only generate a YUV420P picture, we must convert it + * to the codec pixel format if needed */ + if (!sws_ctx) { + sws_ctx = sws_getContext(c->width, c->height, BMP_FORMAT, + c->width, c->height, c->pix_fmt, + sws_flags, NULL, NULL, NULL); + if (!sws_ctx) { + fprintf(stderr, + "Could not initialize the conversion context\n"); + exit(1); } + } + fill_rgba_image(&src_picture, frame_count, c->width, c->height); + sws_scale(sws_ctx, + (const uint8_t * const *)src_picture.data, src_picture.linesize, + 0, c->height, dst_picture.data, dst_picture.linesize); + } else { + fill_rgba_image(&dst_picture, frame_count, c->width, c->height); + } - if (oc->oformat->flags & AVFMT_RAWPICTURE) { - /* Raw video case - directly store the picture in the packet */ - AVPacket pkt; - av_init_packet(&pkt); - - pkt.flags |= AV_PKT_FLAG_KEY; - pkt.stream_index = st->index; - pkt.data = dst_picture.data[0]; - pkt.size = sizeof(AVPicture); - - ret = av_interleaved_write_frame(oc, &pkt); - } else { - AVPacket pkt = { 0 }; - int got_packet; - av_init_packet(&pkt); - - /* encode the image */ - ret = avcodec_encode_video2(c, &pkt, frame, &got_packet); - if (ret < 0) { - fprintf(stderr, "Error encoding video frame: %s\n", av_err2str(ret)); - exit(1); - } - /* If size is zero, it means the image was buffered. */ - - if (!ret && got_packet && pkt.size) { - pkt.stream_index = st->index; - - /* Write the compressed frame to the media file. */ - ret = av_interleaved_write_frame(oc, &pkt); - } else { - ret = 0; - } - } - if (ret != 0) { - fprintf(stderr, "Error while writing video frame: %s\n", av_err2str(ret)); - exit(1); - } - frame_count++; + if (oc->oformat->flags & AVFMT_RAWPICTURE) { + /* Raw video case - directly store the picture in the packet */ + AVPacket pkt; + av_init_packet(&pkt); + + pkt.flags |= AV_PKT_FLAG_KEY; + pkt.stream_index = st->index; + pkt.data = dst_picture.data[0]; + pkt.size = sizeof(AVPicture); + + ret = av_interleaved_write_frame(oc, &pkt); + } else { + AVPacket pkt = { 0 }; + int got_packet; + av_init_packet(&pkt); + + /* encode the image */ + ret = avcodec_encode_video2(c, &pkt, frame, &got_packet); + if (ret < 0) { + fprintf(stderr, "Error encoding video frame: %s\n", av_err2str(ret)); + exit(1); + } + /* If size is zero, it means the image was buffered. */ + + if (!ret && got_packet && pkt.size) { + pkt.stream_index = st->index; + + /* Write the compressed frame to the media file. */ + ret = av_interleaved_write_frame(oc, &pkt); + } else { + ret = 0; + } + } + if (ret != 0) { + fprintf(stderr, "Error while writing video frame: %s\n", av_err2str(ret)); + exit(1); + } + frame_count++; } -static void close_video(AVFormatContext *oc, AVStream *st) -{ - avcodec_close(st->codec); - av_free(src_picture.data[0]); - av_free(dst_picture.data[0]); - av_free(frame); +static void close_video(AVFormatContext *oc, AVStream *st) { + avcodec_close(st->codec); + av_free(src_picture.data[0]); + av_free(dst_picture.data[0]); + av_free(frame); } /**************************************************************/ /* media file output */ -int main(int argc, char **argv) -{ - const char *filename; - AVOutputFormat *fmt; - AVFormatContext *oc; - AVStream *video_st; - AVCodec *video_codec; - int ret; - - /* Initialize libavcodec, and register all codecs and formats. */ - av_register_all(); - - filename = "/Users/sradomski/Desktop/test.mpg"; - - /* allocate the output media context */ - avformat_alloc_output_context2(&oc, NULL, NULL, filename); - if (!oc) { - printf("Could not deduce output format from file extension: using MPEG.\n"); - avformat_alloc_output_context2(&oc, NULL, "mpeg", filename); - } - if (!oc) { - return 1; - } - fmt = oc->oformat; - - /* Add the audio and video streams using the default format codecs - * and initialize the codecs. */ - video_st = NULL; - - if (fmt->video_codec != AV_CODEC_ID_NONE) { - video_st = add_stream(oc, &video_codec, fmt->video_codec); - } - - /* Now that all the parameters are set, we can open the audio and - * video codecs and allocate the necessary encode buffers. */ - if (video_st) - open_video(oc, video_codec, video_st); - - /* open the output file, if needed */ - if (!(fmt->flags & AVFMT_NOFILE)) { - ret = avio_open(&oc->pb, filename, AVIO_FLAG_WRITE); - if (ret < 0) { - fprintf(stderr, "Could not open '%s': %s\n", filename, - av_err2str(ret)); - return 1; - } - } - - /* Write the stream header, if any. */ - ret = avformat_write_header(oc, NULL); - if (ret < 0) { - fprintf(stderr, "Error occurred when opening output file: %s\n", - av_err2str(ret)); - return 1; - } - - if (frame) - frame->pts = 0; - for (int i = 0; i < 125; i++) { - write_video_frame(oc, video_st); - frame->pts += av_rescale_q(1, video_st->codec->time_base, video_st->time_base); - } - - /* Write the trailer, if any. The trailer must be written before you - * close the CodecContexts open when you wrote the header; otherwise - * av_write_trailer() may try to use memory that was freed on - * av_codec_close(). */ - av_write_trailer(oc); - - /* Close each codec. */ - if (video_st) - close_video(oc, video_st); - - if (!(fmt->flags & AVFMT_NOFILE)) - /* Close the output file. */ - avio_close(oc->pb); - - /* free the stream */ - avformat_free_context(oc); - - return 0; +int main(int argc, char **argv) { + const char *filename; + AVOutputFormat *fmt; + AVFormatContext *oc; + AVStream *video_st; + AVCodec *video_codec; + int ret; + + /* Initialize libavcodec, and register all codecs and formats. */ + av_register_all(); + + filename = "/Users/sradomski/Desktop/test.mpg"; + + /* allocate the output media context */ + avformat_alloc_output_context2(&oc, NULL, NULL, filename); + if (!oc) { + printf("Could not deduce output format from file extension: using MPEG.\n"); + avformat_alloc_output_context2(&oc, NULL, "mpeg", filename); + } + if (!oc) { + return 1; + } + fmt = oc->oformat; + + /* Add the audio and video streams using the default format codecs + * and initialize the codecs. */ + video_st = NULL; + + if (fmt->video_codec != AV_CODEC_ID_NONE) { + video_st = add_stream(oc, &video_codec, fmt->video_codec); + } + + /* Now that all the parameters are set, we can open the audio and + * video codecs and allocate the necessary encode buffers. */ + if (video_st) + open_video(oc, video_codec, video_st); + + /* open the output file, if needed */ + if (!(fmt->flags & AVFMT_NOFILE)) { + ret = avio_open(&oc->pb, filename, AVIO_FLAG_WRITE); + if (ret < 0) { + fprintf(stderr, "Could not open '%s': %s\n", filename, + av_err2str(ret)); + return 1; + } + } + + /* Write the stream header, if any. */ + ret = avformat_write_header(oc, NULL); + if (ret < 0) { + fprintf(stderr, "Error occurred when opening output file: %s\n", + av_err2str(ret)); + return 1; + } + + if (frame) + frame->pts = 0; + for (int i = 0; i < 125; i++) { + write_video_frame(oc, video_st); + frame->pts += av_rescale_q(1, video_st->codec->time_base, video_st->time_base); + } + + /* Write the trailer, if any. The trailer must be written before you + * close the CodecContexts open when you wrote the header; otherwise + * av_write_trailer() may try to use memory that was freed on + * av_codec_close(). */ + av_write_trailer(oc); + + /* Close each codec. */ + if (video_st) + close_video(oc, video_st); + + if (!(fmt->flags & AVFMT_NOFILE)) + /* Close the output file. */ + avio_close(oc->pb); + + /* free the stream */ + avformat_free_context(oc); + + return 0; } diff --git a/test/src/test-instant-messaging.cpp b/test/src/test-instant-messaging.cpp index 0bf2898..a63df34 100644 --- a/test/src/test-instant-messaging.cpp +++ b/test/src/test-instant-messaging.cpp @@ -51,13 +51,11 @@ typedef struct _PurpleGLibIOClosure { gpointer data; } PurpleGLibIOClosure; -static void purple_glib_io_destroy(gpointer data) -{ +static void purple_glib_io_destroy(gpointer data) { g_free(data); } -static gboolean purple_glib_io_invoke(GIOChannel *source, GIOCondition condition, gpointer data) -{ +static gboolean purple_glib_io_invoke(GIOChannel *source, GIOCondition condition, gpointer data) { PurpleGLibIOClosure *closure = (PurpleGLibIOClosure*)data; int purple_cond = 0; @@ -67,14 +65,13 @@ static gboolean purple_glib_io_invoke(GIOChannel *source, GIOCondition condition purple_cond |= PURPLE_INPUT_WRITE; closure->function(closure->data, g_io_channel_unix_get_fd(source), - (PurpleInputCondition)purple_cond); + (PurpleInputCondition)purple_cond); return TRUE; } static guint glib_input_add(gint fd, PurpleInputCondition condition, PurpleInputFunction function, - gpointer data) -{ + gpointer data) { PurpleGLibIOClosure *closure = g_new0(PurpleGLibIOClosure, 1); GIOChannel *channel; int cond = 0; @@ -93,14 +90,13 @@ static guint glib_input_add(gint fd, PurpleInputCondition condition, PurpleInput channel = g_io_channel_unix_new(fd); #endif closure->result = g_io_add_watch_full(channel, G_PRIORITY_DEFAULT, (GIOCondition)cond, - purple_glib_io_invoke, closure, purple_glib_io_destroy); + purple_glib_io_invoke, closure, purple_glib_io_destroy); g_io_channel_unref(channel); return closure->result; } -static PurpleEventLoopUiOps glib_eventloops = -{ +static PurpleEventLoopUiOps glib_eventloops = { g_timeout_add, g_source_remove, glib_input_add, @@ -118,8 +114,7 @@ static PurpleEventLoopUiOps glib_eventloops = /*** Conversation uiops ***/ static void null_write_conv(PurpleConversation *conv, const char *who, const char *alias, - const char *message, PurpleMessageFlags flags, time_t mtime) -{ + const char *message, PurpleMessageFlags flags, time_t mtime) { const char *name; if (alias && *alias) name = alias; @@ -129,12 +124,11 @@ null_write_conv(PurpleConversation *conv, const char *who, const char *alias, name = NULL; printf("(%s) %s %s: %s\n", purple_conversation_get_name(conv), - purple_utf8_strftime("(%H:%M:%S)", localtime(&mtime)), - name, message); + purple_utf8_strftime("(%H:%M:%S)", localtime(&mtime)), + name, message); } -static PurpleConversationUiOps null_conv_uiops = -{ +static PurpleConversationUiOps null_conv_uiops = { NULL, /* create_conversation */ NULL, /* destroy_conversation */ NULL, /* write_chat */ @@ -157,8 +151,7 @@ static PurpleConversationUiOps null_conv_uiops = }; static void -null_ui_init(void) -{ +null_ui_init(void) { /** * This should initialize the UI components for all the modules. Here we * just initialize the UI for conversations. @@ -166,8 +159,7 @@ null_ui_init(void) purple_conversations_set_ui_ops(&null_conv_uiops); } -static PurpleCoreUiOps null_core_uiops = -{ +static PurpleCoreUiOps null_core_uiops = { NULL, NULL, null_ui_init, @@ -181,8 +173,7 @@ static PurpleCoreUiOps null_core_uiops = }; static void -init_libpurple(void) -{ +init_libpurple(void) { /* Set a custom user directory (optional) */ purple_util_set_user_dir(CUSTOM_USER_DIRECTORY); @@ -212,8 +203,8 @@ init_libpurple(void) if (!purple_core_init(UI_ID)) { /* Initializing the core failed. Terminate. */ fprintf(stderr, - "libpurple initialization failed. Dumping core.\n" - "Please report this!\n"); + "libpurple initialization failed. Dumping core.\n" + "Please report this!\n"); abort(); } @@ -226,8 +217,7 @@ init_libpurple(void) } static void -signed_on(PurpleConnection *gc, gpointer null) -{ +signed_on(PurpleConnection *gc, gpointer null) { PurpleAccount *account = purple_connection_get_account(gc); printf("Account connected: %s %s\n", purple_account_get_username(account), purple_account_get_protocol_id(account)); } @@ -245,17 +235,15 @@ buddy_signed_on(PurpleBuddy *buddy) { } static void -connect_to_signals_for_demonstration_purposes_only(void) -{ +connect_to_signals_for_demonstration_purposes_only(void) { static int handle; purple_signal_connect(purple_connections_get_handle(), "signed-on", &handle, - PURPLE_CALLBACK(signed_on), NULL); + PURPLE_CALLBACK(signed_on), NULL); purple_signal_connect(purple_blist_get_handle(), "buddy-signed-on", &handle, - PURPLE_CALLBACK(buddy_signed_on), NULL); + PURPLE_CALLBACK(buddy_signed_on), NULL); } -int main(int argc, char *argv[]) -{ +int main(int argc, char *argv[]) { GList *iter; GMainLoop *loop = g_main_loop_new(NULL, FALSE); PurpleAccount *account; diff --git a/test/src/test-url.cpp b/test/src/test-url.cpp index ed892b8..e0f8343 100644 --- a/test/src/test-url.cpp +++ b/test/src/test-url.cpp @@ -47,7 +47,7 @@ int main(int argc, char** argv) { WSADATA wsaData;
WSAStartup(MAKEWORD(2, 2), &wsaData);
#endif
-
+
std::string exeName = argv[0];
exeName = exeName.substr(exeName.find_last_of("\\/") + 1);
diff --git a/test/src/test-w3c.cpp b/test/src/test-w3c.cpp index db8c764..768f6be 100644 --- a/test/src/test-w3c.cpp +++ b/test/src/test-w3c.cpp @@ -139,7 +139,7 @@ int main(int argc, char** argv) { // if (interpreter.getDataModel().getNames().find("ecmascript") != interpreter.getDataModel().getNames().end()) { // } - + interpreter.start(); while(interpreter.runOnMainThread(25)); } |