summaryrefslogtreecommitdiffstats
path: root/test
diff options
context:
space:
mode:
authorStefan Radomski <radomski@tk.informatik.tu-darmstadt.de>2013-10-16 18:35:17 (GMT)
committerStefan Radomski <radomski@tk.informatik.tu-darmstadt.de>2013-10-16 18:35:17 (GMT)
commit60c543c1d8c6699cc96829f15b16ecc60ac0b410 (patch)
treef0087f620d45246acf0268da539cb815796d88c0 /test
parent4d186e24e44fe07f04708c2dc38387534c02c7fd (diff)
downloaduscxml-60c543c1d8c6699cc96829f15b16ecc60ac0b410.zip
uscxml-60c543c1d8c6699cc96829f15b16ecc60ac0b410.tar.gz
uscxml-60c543c1d8c6699cc96829f15b16ecc60ac0b410.tar.bz2
Added license information and bug fixes
Diffstat (limited to 'test')
-rw-r--r--test/CMakeLists.txt6
-rw-r--r--test/src/test-cmdline-parsing.cpp8
-rw-r--r--test/src/test-ffmpeg.cpp518
-rw-r--r--test/src/test-instant-messaging.cpp52
-rw-r--r--test/src/test-url.cpp2
-rw-r--r--test/src/test-w3c.cpp2
6 files changed, 281 insertions, 307 deletions
diff --git a/test/CMakeLists.txt b/test/CMakeLists.txt
index 261817b..eff801c 100644
--- a/test/CMakeLists.txt
+++ b/test/CMakeLists.txt
@@ -118,9 +118,9 @@ set_target_properties(scxml-test-framework-client PROPERTIES FOLDER "Tests")
# add_test(test-curl-multi-api ${CMAKE_RUNTIME_OUTPUT_DIRECTORY}/test-completion)
# set_target_properties(test-curl-multi-api PROPERTIES FOLDER "Tests")
- add_executable(test-w3c src/test-w3c.cpp)
- target_link_libraries(test-w3c uscxml)
- set_target_properties(test-w3c PROPERTIES FOLDER "Tests")
+add_executable(test-w3c src/test-w3c.cpp)
+target_link_libraries(test-w3c uscxml)
+set_target_properties(test-w3c PROPERTIES FOLDER "Tests")
file(GLOB_RECURSE W3C_TESTS
samples/w3c/*.scxml
diff --git a/test/src/test-cmdline-parsing.cpp b/test/src/test-cmdline-parsing.cpp
index 9ff094a..a610248 100644
--- a/test/src/test-cmdline-parsing.cpp
+++ b/test/src/test-cmdline-parsing.cpp
@@ -33,8 +33,6 @@ int main(int argc, char** argv) {
assert(options.logLevel == 10);
assert(!options.withHTTP);
assert(!options); // invalid as no SCXML document is given
-
- optind = 0;
}
if (true) {
@@ -50,8 +48,6 @@ int main(int argc, char** argv) {
assert(options.verbose);
assert(options.interpreters.size() == 1);
assert(options.interpreters.find("/foo/bar.scxml") != options.interpreters.end());
-
- optind = 0;
}
if (true) {
@@ -76,8 +72,6 @@ int main(int argc, char** argv) {
assert(!options.interpreters["/foo/bar1.scxml"]->withHTTP);
assert(options.interpreters["/foo/bar2.scxml"]->withHTTP);
assert(!options.interpreters["/foo/bar3.scxml"]->withHTTP);
-
- optind = 0;
}
if (true) {
@@ -97,8 +91,6 @@ int main(int argc, char** argv) {
assert(options.interpreters.find("/foo/bar1.scxml") != options.interpreters.end());
assert(options.interpreters["/foo/bar1.scxml"]->additionalParameters.find("vrml-path") != options.interpreters["/foo/bar1.scxml"]->additionalParameters.end());
assert(options.interpreters["/foo/bar1.scxml"]->additionalParameters.find("tmp-path") != options.interpreters["/foo/bar1.scxml"]->additionalParameters.end());
-
- optind = 0;
}
return EXIT_SUCCESS;
diff --git a/test/src/test-ffmpeg.cpp b/test/src/test-ffmpeg.cpp
index c9dcfe1..a8660e5 100644
--- a/test/src/test-ffmpeg.cpp
+++ b/test/src/test-ffmpeg.cpp
@@ -53,71 +53,70 @@ static int sws_flags = SWS_BICUBIC;
/* Add an output stream. */
static AVStream *add_stream(AVFormatContext *oc, AVCodec **codec,
- enum AVCodecID codec_id)
-{
- AVCodecContext *c;
- AVStream *st;
-
- /* find the encoder */
- *codec = avcodec_find_encoder(codec_id);
- if (!(*codec)) {
- fprintf(stderr, "Could not find encoder for '%s'\n",
- avcodec_get_name(codec_id));
- exit(1);
- }
-
- st = avformat_new_stream(oc, *codec);
- if (!st) {
- fprintf(stderr, "Could not allocate stream\n");
- exit(1);
- }
- st->id = oc->nb_streams-1;
- c = st->codec;
-
- switch ((*codec)->type) {
- case AVMEDIA_TYPE_AUDIO:
- c->sample_fmt = AV_SAMPLE_FMT_FLTP;
- c->bit_rate = 64000;
- c->sample_rate = 44100;
- c->channels = 2;
- break;
-
- case AVMEDIA_TYPE_VIDEO:
- c->codec_id = codec_id;
-
- c->bit_rate = 400000;
- /* Resolution must be a multiple of two. */
- c->width = 352;
- c->height = 288;
- /* timebase: This is the fundamental unit of time (in seconds) in terms
- * of which frame timestamps are represented. For fixed-fps content,
- * timebase should be 1/framerate and timestamp increments should be
- * identical to 1. */
- c->time_base.den = STREAM_FRAME_RATE;
- c->time_base.num = 1;
- c->gop_size = 12; /* emit one intra frame every twelve frames at most */
- c->pix_fmt = STREAM_PIX_FMT;
- if (c->codec_id == AV_CODEC_ID_MPEG2VIDEO) {
- /* just for testing, we also add B frames */
- c->max_b_frames = 2;
- }
- if (c->codec_id == AV_CODEC_ID_MPEG1VIDEO) {
- /* Needed to avoid using macroblocks in which some coeffs overflow.
- * This does not happen with normal video, it just happens here as
- * the motion of the chroma plane does not match the luma plane. */
- c->mb_decision = 2;
- }
- break;
-
- default:
- break;
- }
-
- /* Some formats want stream headers to be separate. */
- if (oc->oformat->flags & AVFMT_GLOBALHEADER)
- c->flags |= CODEC_FLAG_GLOBAL_HEADER;
-
- return st;
+ enum AVCodecID codec_id) {
+ AVCodecContext *c;
+ AVStream *st;
+
+ /* find the encoder */
+ *codec = avcodec_find_encoder(codec_id);
+ if (!(*codec)) {
+ fprintf(stderr, "Could not find encoder for '%s'\n",
+ avcodec_get_name(codec_id));
+ exit(1);
+ }
+
+ st = avformat_new_stream(oc, *codec);
+ if (!st) {
+ fprintf(stderr, "Could not allocate stream\n");
+ exit(1);
+ }
+ st->id = oc->nb_streams-1;
+ c = st->codec;
+
+ switch ((*codec)->type) {
+ case AVMEDIA_TYPE_AUDIO:
+ c->sample_fmt = AV_SAMPLE_FMT_FLTP;
+ c->bit_rate = 64000;
+ c->sample_rate = 44100;
+ c->channels = 2;
+ break;
+
+ case AVMEDIA_TYPE_VIDEO:
+ c->codec_id = codec_id;
+
+ c->bit_rate = 400000;
+ /* Resolution must be a multiple of two. */
+ c->width = 352;
+ c->height = 288;
+ /* timebase: This is the fundamental unit of time (in seconds) in terms
+ * of which frame timestamps are represented. For fixed-fps content,
+ * timebase should be 1/framerate and timestamp increments should be
+ * identical to 1. */
+ c->time_base.den = STREAM_FRAME_RATE;
+ c->time_base.num = 1;
+ c->gop_size = 12; /* emit one intra frame every twelve frames at most */
+ c->pix_fmt = STREAM_PIX_FMT;
+ if (c->codec_id == AV_CODEC_ID_MPEG2VIDEO) {
+ /* just for testing, we also add B frames */
+ c->max_b_frames = 2;
+ }
+ if (c->codec_id == AV_CODEC_ID_MPEG1VIDEO) {
+ /* Needed to avoid using macroblocks in which some coeffs overflow.
+ * This does not happen with normal video, it just happens here as
+ * the motion of the chroma plane does not match the luma plane. */
+ c->mb_decision = 2;
+ }
+ break;
+
+ default:
+ break;
+ }
+
+ /* Some formats want stream headers to be separate. */
+ if (oc->oformat->flags & AVFMT_GLOBALHEADER)
+ c->flags |= CODEC_FLAG_GLOBAL_HEADER;
+
+ return st;
}
@@ -128,60 +127,58 @@ static AVFrame *frame;
static AVPicture src_picture, dst_picture;
static int frame_count;
-static void open_video(AVFormatContext *oc, AVCodec *codec, AVStream *st)
-{
- int ret;
- AVCodecContext *c = st->codec;
-
- /* open the codec */
- ret = avcodec_open2(c, codec, NULL);
- if (ret < 0) {
- fprintf(stderr, "Could not open video codec: %s\n", av_err2str(ret));
- exit(1);
- }
-
- /* allocate and init a re-usable frame */
- frame = avcodec_alloc_frame();
- if (!frame) {
- fprintf(stderr, "Could not allocate video frame\n");
- exit(1);
- }
-
- /* Allocate the encoded raw picture. */
- ret = avpicture_alloc(&dst_picture, c->pix_fmt, c->width, c->height);
- if (ret < 0) {
- fprintf(stderr, "Could not allocate picture: %s\n", av_err2str(ret));
- exit(1);
- }
-
- /* If the output format is not YUV420P, then a temporary YUV420P
- * picture is needed too. It is then converted to the required
- * output format. */
- if (c->pix_fmt != BMP_FORMAT) {
- ret = avpicture_alloc(&src_picture, BMP_FORMAT, c->width, c->height);
- if (ret < 0) {
- fprintf(stderr, "Could not allocate temporary picture: %s\n",
- av_err2str(ret));
- exit(1);
- }
- }
-
- /* copy data and linesize picture pointers to frame */
- *((AVPicture *)frame) = dst_picture;
+static void open_video(AVFormatContext *oc, AVCodec *codec, AVStream *st) {
+ int ret;
+ AVCodecContext *c = st->codec;
+
+ /* open the codec */
+ ret = avcodec_open2(c, codec, NULL);
+ if (ret < 0) {
+ fprintf(stderr, "Could not open video codec: %s\n", av_err2str(ret));
+ exit(1);
+ }
+
+ /* allocate and init a re-usable frame */
+ frame = avcodec_alloc_frame();
+ if (!frame) {
+ fprintf(stderr, "Could not allocate video frame\n");
+ exit(1);
+ }
+
+ /* Allocate the encoded raw picture. */
+ ret = avpicture_alloc(&dst_picture, c->pix_fmt, c->width, c->height);
+ if (ret < 0) {
+ fprintf(stderr, "Could not allocate picture: %s\n", av_err2str(ret));
+ exit(1);
+ }
+
+ /* If the output format is not YUV420P, then a temporary YUV420P
+ * picture is needed too. It is then converted to the required
+ * output format. */
+ if (c->pix_fmt != BMP_FORMAT) {
+ ret = avpicture_alloc(&src_picture, BMP_FORMAT, c->width, c->height);
+ if (ret < 0) {
+ fprintf(stderr, "Could not allocate temporary picture: %s\n",
+ av_err2str(ret));
+ exit(1);
+ }
+ }
+
+ /* copy data and linesize picture pointers to frame */
+ *((AVPicture *)frame) = dst_picture;
}
/* Prepare a dummy image. */
static void fill_rgba_image(AVPicture *pict, int frame_index,
- int width, int height)
-{
+ int width, int height) {
int x, y, i;
-
+
i = frame_index;
-
+
std::stringstream ssFilename;
ssFilename << "/Users/sradomski/Desktop/ctrl/" << (i % 125) << ".bmp";
-
+
std::ifstream file(ssFilename.str().c_str());
file.seekg(0, std::ios::end);
@@ -190,7 +187,7 @@ static void fill_rgba_image(AVPicture *pict, int frame_index,
char* buffer = (char*)malloc(length);
file.read(buffer, length);
-
+
uint32_t offset = 0;
offset += buffer[10] << 0;
offset += buffer[11] << 8;
@@ -205,167 +202,164 @@ static void fill_rgba_image(AVPicture *pict, int frame_index,
pict->data[0][y * pict->linesize[0] + x * 3 + 2] = buffer[offset++];
}
}
-
+
free(buffer);
}
-static void write_video_frame(AVFormatContext *oc, AVStream *st)
-{
- int ret;
- static struct SwsContext *sws_ctx;
- AVCodecContext *c = st->codec;
-
- if (c->pix_fmt != BMP_FORMAT) {
- /* as we only generate a YUV420P picture, we must convert it
- * to the codec pixel format if needed */
- if (!sws_ctx) {
- sws_ctx = sws_getContext(c->width, c->height, BMP_FORMAT,
- c->width, c->height, c->pix_fmt,
- sws_flags, NULL, NULL, NULL);
- if (!sws_ctx) {
- fprintf(stderr,
- "Could not initialize the conversion context\n");
- exit(1);
- }
- }
- fill_rgba_image(&src_picture, frame_count, c->width, c->height);
- sws_scale(sws_ctx,
- (const uint8_t * const *)src_picture.data, src_picture.linesize,
- 0, c->height, dst_picture.data, dst_picture.linesize);
- } else {
- fill_rgba_image(&dst_picture, frame_count, c->width, c->height);
+static void write_video_frame(AVFormatContext *oc, AVStream *st) {
+ int ret;
+ static struct SwsContext *sws_ctx;
+ AVCodecContext *c = st->codec;
+
+ if (c->pix_fmt != BMP_FORMAT) {
+ /* as we only generate a YUV420P picture, we must convert it
+ * to the codec pixel format if needed */
+ if (!sws_ctx) {
+ sws_ctx = sws_getContext(c->width, c->height, BMP_FORMAT,
+ c->width, c->height, c->pix_fmt,
+ sws_flags, NULL, NULL, NULL);
+ if (!sws_ctx) {
+ fprintf(stderr,
+ "Could not initialize the conversion context\n");
+ exit(1);
}
+ }
+ fill_rgba_image(&src_picture, frame_count, c->width, c->height);
+ sws_scale(sws_ctx,
+ (const uint8_t * const *)src_picture.data, src_picture.linesize,
+ 0, c->height, dst_picture.data, dst_picture.linesize);
+ } else {
+ fill_rgba_image(&dst_picture, frame_count, c->width, c->height);
+ }
- if (oc->oformat->flags & AVFMT_RAWPICTURE) {
- /* Raw video case - directly store the picture in the packet */
- AVPacket pkt;
- av_init_packet(&pkt);
-
- pkt.flags |= AV_PKT_FLAG_KEY;
- pkt.stream_index = st->index;
- pkt.data = dst_picture.data[0];
- pkt.size = sizeof(AVPicture);
-
- ret = av_interleaved_write_frame(oc, &pkt);
- } else {
- AVPacket pkt = { 0 };
- int got_packet;
- av_init_packet(&pkt);
-
- /* encode the image */
- ret = avcodec_encode_video2(c, &pkt, frame, &got_packet);
- if (ret < 0) {
- fprintf(stderr, "Error encoding video frame: %s\n", av_err2str(ret));
- exit(1);
- }
- /* If size is zero, it means the image was buffered. */
-
- if (!ret && got_packet && pkt.size) {
- pkt.stream_index = st->index;
-
- /* Write the compressed frame to the media file. */
- ret = av_interleaved_write_frame(oc, &pkt);
- } else {
- ret = 0;
- }
- }
- if (ret != 0) {
- fprintf(stderr, "Error while writing video frame: %s\n", av_err2str(ret));
- exit(1);
- }
- frame_count++;
+ if (oc->oformat->flags & AVFMT_RAWPICTURE) {
+ /* Raw video case - directly store the picture in the packet */
+ AVPacket pkt;
+ av_init_packet(&pkt);
+
+ pkt.flags |= AV_PKT_FLAG_KEY;
+ pkt.stream_index = st->index;
+ pkt.data = dst_picture.data[0];
+ pkt.size = sizeof(AVPicture);
+
+ ret = av_interleaved_write_frame(oc, &pkt);
+ } else {
+ AVPacket pkt = { 0 };
+ int got_packet;
+ av_init_packet(&pkt);
+
+ /* encode the image */
+ ret = avcodec_encode_video2(c, &pkt, frame, &got_packet);
+ if (ret < 0) {
+ fprintf(stderr, "Error encoding video frame: %s\n", av_err2str(ret));
+ exit(1);
+ }
+ /* If size is zero, it means the image was buffered. */
+
+ if (!ret && got_packet && pkt.size) {
+ pkt.stream_index = st->index;
+
+ /* Write the compressed frame to the media file. */
+ ret = av_interleaved_write_frame(oc, &pkt);
+ } else {
+ ret = 0;
+ }
+ }
+ if (ret != 0) {
+ fprintf(stderr, "Error while writing video frame: %s\n", av_err2str(ret));
+ exit(1);
+ }
+ frame_count++;
}
-static void close_video(AVFormatContext *oc, AVStream *st)
-{
- avcodec_close(st->codec);
- av_free(src_picture.data[0]);
- av_free(dst_picture.data[0]);
- av_free(frame);
+static void close_video(AVFormatContext *oc, AVStream *st) {
+ avcodec_close(st->codec);
+ av_free(src_picture.data[0]);
+ av_free(dst_picture.data[0]);
+ av_free(frame);
}
/**************************************************************/
/* media file output */
-int main(int argc, char **argv)
-{
- const char *filename;
- AVOutputFormat *fmt;
- AVFormatContext *oc;
- AVStream *video_st;
- AVCodec *video_codec;
- int ret;
-
- /* Initialize libavcodec, and register all codecs and formats. */
- av_register_all();
-
- filename = "/Users/sradomski/Desktop/test.mpg";
-
- /* allocate the output media context */
- avformat_alloc_output_context2(&oc, NULL, NULL, filename);
- if (!oc) {
- printf("Could not deduce output format from file extension: using MPEG.\n");
- avformat_alloc_output_context2(&oc, NULL, "mpeg", filename);
- }
- if (!oc) {
- return 1;
- }
- fmt = oc->oformat;
-
- /* Add the audio and video streams using the default format codecs
- * and initialize the codecs. */
- video_st = NULL;
-
- if (fmt->video_codec != AV_CODEC_ID_NONE) {
- video_st = add_stream(oc, &video_codec, fmt->video_codec);
- }
-
- /* Now that all the parameters are set, we can open the audio and
- * video codecs and allocate the necessary encode buffers. */
- if (video_st)
- open_video(oc, video_codec, video_st);
-
- /* open the output file, if needed */
- if (!(fmt->flags & AVFMT_NOFILE)) {
- ret = avio_open(&oc->pb, filename, AVIO_FLAG_WRITE);
- if (ret < 0) {
- fprintf(stderr, "Could not open '%s': %s\n", filename,
- av_err2str(ret));
- return 1;
- }
- }
-
- /* Write the stream header, if any. */
- ret = avformat_write_header(oc, NULL);
- if (ret < 0) {
- fprintf(stderr, "Error occurred when opening output file: %s\n",
- av_err2str(ret));
- return 1;
- }
-
- if (frame)
- frame->pts = 0;
- for (int i = 0; i < 125; i++) {
- write_video_frame(oc, video_st);
- frame->pts += av_rescale_q(1, video_st->codec->time_base, video_st->time_base);
- }
-
- /* Write the trailer, if any. The trailer must be written before you
- * close the CodecContexts open when you wrote the header; otherwise
- * av_write_trailer() may try to use memory that was freed on
- * av_codec_close(). */
- av_write_trailer(oc);
-
- /* Close each codec. */
- if (video_st)
- close_video(oc, video_st);
-
- if (!(fmt->flags & AVFMT_NOFILE))
- /* Close the output file. */
- avio_close(oc->pb);
-
- /* free the stream */
- avformat_free_context(oc);
-
- return 0;
+int main(int argc, char **argv) {
+ const char *filename;
+ AVOutputFormat *fmt;
+ AVFormatContext *oc;
+ AVStream *video_st;
+ AVCodec *video_codec;
+ int ret;
+
+ /* Initialize libavcodec, and register all codecs and formats. */
+ av_register_all();
+
+ filename = "/Users/sradomski/Desktop/test.mpg";
+
+ /* allocate the output media context */
+ avformat_alloc_output_context2(&oc, NULL, NULL, filename);
+ if (!oc) {
+ printf("Could not deduce output format from file extension: using MPEG.\n");
+ avformat_alloc_output_context2(&oc, NULL, "mpeg", filename);
+ }
+ if (!oc) {
+ return 1;
+ }
+ fmt = oc->oformat;
+
+ /* Add the audio and video streams using the default format codecs
+ * and initialize the codecs. */
+ video_st = NULL;
+
+ if (fmt->video_codec != AV_CODEC_ID_NONE) {
+ video_st = add_stream(oc, &video_codec, fmt->video_codec);
+ }
+
+ /* Now that all the parameters are set, we can open the audio and
+ * video codecs and allocate the necessary encode buffers. */
+ if (video_st)
+ open_video(oc, video_codec, video_st);
+
+ /* open the output file, if needed */
+ if (!(fmt->flags & AVFMT_NOFILE)) {
+ ret = avio_open(&oc->pb, filename, AVIO_FLAG_WRITE);
+ if (ret < 0) {
+ fprintf(stderr, "Could not open '%s': %s\n", filename,
+ av_err2str(ret));
+ return 1;
+ }
+ }
+
+ /* Write the stream header, if any. */
+ ret = avformat_write_header(oc, NULL);
+ if (ret < 0) {
+ fprintf(stderr, "Error occurred when opening output file: %s\n",
+ av_err2str(ret));
+ return 1;
+ }
+
+ if (frame)
+ frame->pts = 0;
+ for (int i = 0; i < 125; i++) {
+ write_video_frame(oc, video_st);
+ frame->pts += av_rescale_q(1, video_st->codec->time_base, video_st->time_base);
+ }
+
+ /* Write the trailer, if any. The trailer must be written before you
+ * close the CodecContexts open when you wrote the header; otherwise
+ * av_write_trailer() may try to use memory that was freed on
+ * av_codec_close(). */
+ av_write_trailer(oc);
+
+ /* Close each codec. */
+ if (video_st)
+ close_video(oc, video_st);
+
+ if (!(fmt->flags & AVFMT_NOFILE))
+ /* Close the output file. */
+ avio_close(oc->pb);
+
+ /* free the stream */
+ avformat_free_context(oc);
+
+ return 0;
}
diff --git a/test/src/test-instant-messaging.cpp b/test/src/test-instant-messaging.cpp
index 0bf2898..a63df34 100644
--- a/test/src/test-instant-messaging.cpp
+++ b/test/src/test-instant-messaging.cpp
@@ -51,13 +51,11 @@ typedef struct _PurpleGLibIOClosure {
gpointer data;
} PurpleGLibIOClosure;
-static void purple_glib_io_destroy(gpointer data)
-{
+static void purple_glib_io_destroy(gpointer data) {
g_free(data);
}
-static gboolean purple_glib_io_invoke(GIOChannel *source, GIOCondition condition, gpointer data)
-{
+static gboolean purple_glib_io_invoke(GIOChannel *source, GIOCondition condition, gpointer data) {
PurpleGLibIOClosure *closure = (PurpleGLibIOClosure*)data;
int purple_cond = 0;
@@ -67,14 +65,13 @@ static gboolean purple_glib_io_invoke(GIOChannel *source, GIOCondition condition
purple_cond |= PURPLE_INPUT_WRITE;
closure->function(closure->data, g_io_channel_unix_get_fd(source),
- (PurpleInputCondition)purple_cond);
+ (PurpleInputCondition)purple_cond);
return TRUE;
}
static guint glib_input_add(gint fd, PurpleInputCondition condition, PurpleInputFunction function,
- gpointer data)
-{
+ gpointer data) {
PurpleGLibIOClosure *closure = g_new0(PurpleGLibIOClosure, 1);
GIOChannel *channel;
int cond = 0;
@@ -93,14 +90,13 @@ static guint glib_input_add(gint fd, PurpleInputCondition condition, PurpleInput
channel = g_io_channel_unix_new(fd);
#endif
closure->result = g_io_add_watch_full(channel, G_PRIORITY_DEFAULT, (GIOCondition)cond,
- purple_glib_io_invoke, closure, purple_glib_io_destroy);
+ purple_glib_io_invoke, closure, purple_glib_io_destroy);
g_io_channel_unref(channel);
return closure->result;
}
-static PurpleEventLoopUiOps glib_eventloops =
-{
+static PurpleEventLoopUiOps glib_eventloops = {
g_timeout_add,
g_source_remove,
glib_input_add,
@@ -118,8 +114,7 @@ static PurpleEventLoopUiOps glib_eventloops =
/*** Conversation uiops ***/
static void
null_write_conv(PurpleConversation *conv, const char *who, const char *alias,
- const char *message, PurpleMessageFlags flags, time_t mtime)
-{
+ const char *message, PurpleMessageFlags flags, time_t mtime) {
const char *name;
if (alias && *alias)
name = alias;
@@ -129,12 +124,11 @@ null_write_conv(PurpleConversation *conv, const char *who, const char *alias,
name = NULL;
printf("(%s) %s %s: %s\n", purple_conversation_get_name(conv),
- purple_utf8_strftime("(%H:%M:%S)", localtime(&mtime)),
- name, message);
+ purple_utf8_strftime("(%H:%M:%S)", localtime(&mtime)),
+ name, message);
}
-static PurpleConversationUiOps null_conv_uiops =
-{
+static PurpleConversationUiOps null_conv_uiops = {
NULL, /* create_conversation */
NULL, /* destroy_conversation */
NULL, /* write_chat */
@@ -157,8 +151,7 @@ static PurpleConversationUiOps null_conv_uiops =
};
static void
-null_ui_init(void)
-{
+null_ui_init(void) {
/**
* This should initialize the UI components for all the modules. Here we
* just initialize the UI for conversations.
@@ -166,8 +159,7 @@ null_ui_init(void)
purple_conversations_set_ui_ops(&null_conv_uiops);
}
-static PurpleCoreUiOps null_core_uiops =
-{
+static PurpleCoreUiOps null_core_uiops = {
NULL,
NULL,
null_ui_init,
@@ -181,8 +173,7 @@ static PurpleCoreUiOps null_core_uiops =
};
static void
-init_libpurple(void)
-{
+init_libpurple(void) {
/* Set a custom user directory (optional) */
purple_util_set_user_dir(CUSTOM_USER_DIRECTORY);
@@ -212,8 +203,8 @@ init_libpurple(void)
if (!purple_core_init(UI_ID)) {
/* Initializing the core failed. Terminate. */
fprintf(stderr,
- "libpurple initialization failed. Dumping core.\n"
- "Please report this!\n");
+ "libpurple initialization failed. Dumping core.\n"
+ "Please report this!\n");
abort();
}
@@ -226,8 +217,7 @@ init_libpurple(void)
}
static void
-signed_on(PurpleConnection *gc, gpointer null)
-{
+signed_on(PurpleConnection *gc, gpointer null) {
PurpleAccount *account = purple_connection_get_account(gc);
printf("Account connected: %s %s\n", purple_account_get_username(account), purple_account_get_protocol_id(account));
}
@@ -245,17 +235,15 @@ buddy_signed_on(PurpleBuddy *buddy) {
}
static void
-connect_to_signals_for_demonstration_purposes_only(void)
-{
+connect_to_signals_for_demonstration_purposes_only(void) {
static int handle;
purple_signal_connect(purple_connections_get_handle(), "signed-on", &handle,
- PURPLE_CALLBACK(signed_on), NULL);
+ PURPLE_CALLBACK(signed_on), NULL);
purple_signal_connect(purple_blist_get_handle(), "buddy-signed-on", &handle,
- PURPLE_CALLBACK(buddy_signed_on), NULL);
+ PURPLE_CALLBACK(buddy_signed_on), NULL);
}
-int main(int argc, char *argv[])
-{
+int main(int argc, char *argv[]) {
GList *iter;
GMainLoop *loop = g_main_loop_new(NULL, FALSE);
PurpleAccount *account;
diff --git a/test/src/test-url.cpp b/test/src/test-url.cpp
index ed892b8..e0f8343 100644
--- a/test/src/test-url.cpp
+++ b/test/src/test-url.cpp
@@ -47,7 +47,7 @@ int main(int argc, char** argv) {
WSADATA wsaData;
WSAStartup(MAKEWORD(2, 2), &wsaData);
#endif
-
+
std::string exeName = argv[0];
exeName = exeName.substr(exeName.find_last_of("\\/") + 1);
diff --git a/test/src/test-w3c.cpp b/test/src/test-w3c.cpp
index db8c764..768f6be 100644
--- a/test/src/test-w3c.cpp
+++ b/test/src/test-w3c.cpp
@@ -139,7 +139,7 @@ int main(int argc, char** argv) {
// if (interpreter.getDataModel().getNames().find("ecmascript") != interpreter.getDataModel().getNames().end()) {
// }
-
+
interpreter.start();
while(interpreter.runOnMainThread(25));
}