OLD | NEW |
---|---|
1 // Copyright 2015 The Chromium Authors. All rights reserved. | 1 // Copyright 2015 The Chromium Authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #include <stddef.h> | 5 #include <stddef.h> |
6 #include <stdint.h> | 6 #include <stdint.h> |
7 | 7 |
8 #include <cstdlib> | 8 #include <cstdlib> |
9 #include <memory> | 9 #include <memory> |
10 #include <utility> | 10 #include <utility> |
(...skipping 356 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
367 Bind(&SaveProfilesAndSignalEvent, Unretained(profiles), | 367 Bind(&SaveProfilesAndSignalEvent, Unretained(profiles), |
368 Unretained(&sampling_thread_completed)); | 368 Unretained(&sampling_thread_completed)); |
369 StackSamplingProfiler profiler(target_thread_id, params, callback); | 369 StackSamplingProfiler profiler(target_thread_id, params, callback); |
370 profiler.Start(); | 370 profiler.Start(); |
371 sampling_thread_completed.TimedWait(profiler_wait_time); | 371 sampling_thread_completed.TimedWait(profiler_wait_time); |
372 profiler.Stop(); | 372 profiler.Stop(); |
373 sampling_thread_completed.Wait(); | 373 sampling_thread_completed.Wait(); |
374 }); | 374 }); |
375 } | 375 } |
376 | 376 |
377 // Waits for one of multiple samplings to complete. | |
378 size_t WaitForSamplingComplete( | |
379 std::vector<std::unique_ptr<WaitableEvent>>* sampling_completed) { | |
380 // Map unique_ptrs to something that WaitMany can accept. | |
381 std::vector<WaitableEvent*> sampling_completed_rawptrs( | |
382 sampling_completed->size()); | |
383 std::transform( | |
384 sampling_completed->begin(), sampling_completed->end(), | |
385 sampling_completed_rawptrs.begin(), | |
386 [](const std::unique_ptr<WaitableEvent>& elem) { return elem.get(); }); | |
387 // Wait for one profiler to finish. | |
388 return WaitableEvent::WaitMany(sampling_completed_rawptrs.data(), | |
389 sampling_completed_rawptrs.size()); | |
390 } | |
391 | |
377 // If this executable was linked with /INCREMENTAL (the default for non-official | 392 // If this executable was linked with /INCREMENTAL (the default for non-official |
378 // debug and release builds on Windows), function addresses do not correspond to | 393 // debug and release builds on Windows), function addresses do not correspond to |
379 // function code itself, but instead to instructions in the Incremental Link | 394 // function code itself, but instead to instructions in the Incremental Link |
380 // Table that jump to the functions. Checks for a jump instruction and if | 395 // Table that jump to the functions. Checks for a jump instruction and if |
381 // present does a little decompilation to find the function's actual starting | 396 // present does a little decompilation to find the function's actual starting |
382 // address. | 397 // address. |
383 const void* MaybeFixupFunctionAddressForILT(const void* function_address) { | 398 const void* MaybeFixupFunctionAddressForILT(const void* function_address) { |
384 #if defined(_WIN64) | 399 #if defined(_WIN64) |
385 const unsigned char* opcode = | 400 const unsigned char* opcode = |
386 reinterpret_cast<const unsigned char*>(function_address); | 401 reinterpret_cast<const unsigned char*>(function_address); |
(...skipping 50 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
437 // asynchronous library loading has completed before walking the stack. If | 452 // asynchronous library loading has completed before walking the stack. If |
438 // false, the unloading may still be occurring during the stack walk. | 453 // false, the unloading may still be occurring during the stack walk. |
439 void TestLibraryUnload(bool wait_until_unloaded) { | 454 void TestLibraryUnload(bool wait_until_unloaded) { |
440 // Test delegate that supports intervening between the copying of the stack | 455 // Test delegate that supports intervening between the copying of the stack |
441 // and the walking of the stack. | 456 // and the walking of the stack. |
442 class StackCopiedSignaler : public NativeStackSamplerTestDelegate { | 457 class StackCopiedSignaler : public NativeStackSamplerTestDelegate { |
443 public: | 458 public: |
444 StackCopiedSignaler(WaitableEvent* stack_copied, | 459 StackCopiedSignaler(WaitableEvent* stack_copied, |
445 WaitableEvent* start_stack_walk, | 460 WaitableEvent* start_stack_walk, |
446 bool wait_to_walk_stack) | 461 bool wait_to_walk_stack) |
447 : stack_copied_(stack_copied), start_stack_walk_(start_stack_walk), | 462 : stack_copied_(stack_copied), |
448 wait_to_walk_stack_(wait_to_walk_stack) { | 463 start_stack_walk_(start_stack_walk), |
449 } | 464 wait_to_walk_stack_(wait_to_walk_stack) {} |
450 | 465 |
451 void OnPreStackWalk() override { | 466 void OnPreStackWalk() override { |
452 stack_copied_->Signal(); | 467 stack_copied_->Signal(); |
453 if (wait_to_walk_stack_) | 468 if (wait_to_walk_stack_) |
454 start_stack_walk_->Wait(); | 469 start_stack_walk_->Wait(); |
455 } | 470 } |
456 | 471 |
457 private: | 472 private: |
458 WaitableEvent* const stack_copied_; | 473 WaitableEvent* const stack_copied_; |
459 WaitableEvent* const start_stack_walk_; | 474 WaitableEvent* const start_stack_walk_; |
(...skipping 253 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
713 reinterpret_cast<const void*>(&TargetThread::CallWithAlloca)) | 728 reinterpret_cast<const void*>(&TargetThread::CallWithAlloca)) |
714 << " was not found in stack:\n" | 729 << " was not found in stack:\n" |
715 << FormatSampleForDiagnosticOutput(sample, profile.modules); | 730 << FormatSampleForDiagnosticOutput(sample, profile.modules); |
716 | 731 |
717 // These frames should be adjacent on the stack. | 732 // These frames should be adjacent on the stack. |
718 EXPECT_EQ(1, alloca_frame - end_frame) | 733 EXPECT_EQ(1, alloca_frame - end_frame) |
719 << "Stack:\n" | 734 << "Stack:\n" |
720 << FormatSampleForDiagnosticOutput(sample, profile.modules); | 735 << FormatSampleForDiagnosticOutput(sample, profile.modules); |
721 } | 736 } |
722 | 737 |
723 // Checks that the fire-and-forget interface works. | |
724 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) | |
725 #define MAYBE_StartAndRunAsync StartAndRunAsync | |
726 #else | |
727 #define MAYBE_StartAndRunAsync DISABLED_StartAndRunAsync | |
728 #endif | |
729 TEST(StackSamplingProfilerTest, MAYBE_StartAndRunAsync) { | |
730 // StartAndRunAsync requires the caller to have a message loop. | |
731 MessageLoop message_loop; | |
732 | |
733 SamplingParams params; | |
734 params.samples_per_burst = 1; | |
735 | |
736 CallStackProfiles profiles; | |
737 WithTargetThread([¶ms, &profiles](PlatformThreadId target_thread_id) { | |
738 WaitableEvent sampling_thread_completed( | |
739 WaitableEvent::ResetPolicy::AUTOMATIC, | |
740 WaitableEvent::InitialState::NOT_SIGNALED); | |
741 const StackSamplingProfiler::CompletedCallback callback = | |
742 Bind(&SaveProfilesAndSignalEvent, Unretained(&profiles), | |
743 Unretained(&sampling_thread_completed)); | |
744 StackSamplingProfiler::StartAndRunAsync(target_thread_id, params, callback); | |
745 RunLoop().RunUntilIdle(); | |
746 sampling_thread_completed.Wait(); | |
747 }); | |
748 | |
749 ASSERT_EQ(1u, profiles.size()); | |
750 } | |
751 | |
752 // Checks that the expected number of profiles and samples are present in the | 738 // Checks that the expected number of profiles and samples are present in the |
753 // call stack profiles produced. | 739 // call stack profiles produced. |
754 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) | 740 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) |
755 #define MAYBE_MultipleProfilesAndSamples MultipleProfilesAndSamples | 741 #define MAYBE_MultipleProfilesAndSamples MultipleProfilesAndSamples |
756 #else | 742 #else |
757 #define MAYBE_MultipleProfilesAndSamples DISABLED_MultipleProfilesAndSamples | 743 #define MAYBE_MultipleProfilesAndSamples DISABLED_MultipleProfilesAndSamples |
758 #endif | 744 #endif |
759 TEST(StackSamplingProfilerTest, MAYBE_MultipleProfilesAndSamples) { | 745 TEST(StackSamplingProfilerTest, MAYBE_MultipleProfilesAndSamples) { |
760 SamplingParams params; | 746 SamplingParams params; |
761 params.burst_interval = params.sampling_interval = | 747 params.burst_interval = params.sampling_interval = |
(...skipping 91 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
853 }); | 839 }); |
854 } | 840 } |
855 | 841 |
856 // Checks that the same profiler may be run multiple times. | 842 // Checks that the same profiler may be run multiple times. |
857 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) | 843 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) |
858 #define MAYBE_CanRunMultipleTimes CanRunMultipleTimes | 844 #define MAYBE_CanRunMultipleTimes CanRunMultipleTimes |
859 #else | 845 #else |
860 #define MAYBE_CanRunMultipleTimes DISABLED_CanRunMultipleTimes | 846 #define MAYBE_CanRunMultipleTimes DISABLED_CanRunMultipleTimes |
861 #endif | 847 #endif |
862 TEST(StackSamplingProfilerTest, MAYBE_CanRunMultipleTimes) { | 848 TEST(StackSamplingProfilerTest, MAYBE_CanRunMultipleTimes) { |
849 StackSamplingProfiler::TestAPI::DisableIdleShutdown(); | |
850 | |
851 WithTargetThread([](PlatformThreadId target_thread_id) { | |
852 SamplingParams params; | |
853 params.sampling_interval = TimeDelta::FromMilliseconds(0); | |
854 params.samples_per_burst = 1; | |
855 | |
856 CallStackProfiles profiles; | |
857 WaitableEvent sampling_completed(WaitableEvent::ResetPolicy::MANUAL, | |
858 WaitableEvent::InitialState::NOT_SIGNALED); | |
859 const StackSamplingProfiler::CompletedCallback callback = | |
860 Bind(&SaveProfilesAndSignalEvent, Unretained(&profiles), | |
861 Unretained(&sampling_completed)); | |
862 StackSamplingProfiler profiler(target_thread_id, params, callback); | |
863 | |
864 // Just start and stop to execute code paths. | |
865 profiler.Start(); | |
866 profiler.Stop(); | |
867 sampling_completed.Wait(); | |
868 | |
869 // Ensure a second request will run and not block. | |
870 sampling_completed.Reset(); | |
871 profiles.clear(); | |
872 profiler.Start(); | |
873 sampling_completed.Wait(); | |
874 profiler.Stop(); | |
875 ASSERT_EQ(1u, profiles.size()); | |
876 }); | |
877 } | |
878 | |
879 // Checks that the different profilers may be run. | |
880 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) | |
881 #define MAYBE_CanRunMultipleProfilers CanRunMultipleProfilers | |
882 #else | |
883 #define MAYBE_CanRunMultipleProfilers DISABLED_CanRunMultipleProfilers | |
884 #endif | |
885 TEST(StackSamplingProfilerTest, MAYBE_CanRunMultipleProfilers) { | |
886 StackSamplingProfiler::TestAPI::DisableIdleShutdown(); | |
887 | |
863 SamplingParams params; | 888 SamplingParams params; |
864 params.sampling_interval = TimeDelta::FromMilliseconds(0); | 889 params.sampling_interval = TimeDelta::FromMilliseconds(0); |
865 params.samples_per_burst = 1; | 890 params.samples_per_burst = 1; |
866 | 891 |
867 std::vector<CallStackProfile> profiles; | 892 std::vector<CallStackProfile> profiles; |
868 CaptureProfiles(params, AVeryLongTimeDelta(), &profiles); | 893 CaptureProfiles(params, AVeryLongTimeDelta(), &profiles); |
869 ASSERT_EQ(1u, profiles.size()); | 894 ASSERT_EQ(1u, profiles.size()); |
870 | 895 |
871 profiles.clear(); | 896 profiles.clear(); |
872 CaptureProfiles(params, AVeryLongTimeDelta(), &profiles); | 897 CaptureProfiles(params, AVeryLongTimeDelta(), &profiles); |
873 ASSERT_EQ(1u, profiles.size()); | 898 ASSERT_EQ(1u, profiles.size()); |
874 } | 899 } |
875 | 900 |
876 // Checks that requests to start profiling while another profile is taking place | 901 // Checks that additional requests will restart a stopped profiler. |
877 // are ignored. | |
878 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) | 902 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) |
879 #define MAYBE_ConcurrentProfiling ConcurrentProfiling | 903 #define MAYBE_WillRestartSampler WillRestartSampler |
880 #else | 904 #else |
881 #define MAYBE_ConcurrentProfiling DISABLED_ConcurrentProfiling | 905 #define MAYBE_WillRestartSampler DISABLED_WillRestartSampler |
882 #endif | 906 #endif |
883 TEST(StackSamplingProfilerTest, MAYBE_ConcurrentProfiling) { | 907 TEST(StackSamplingProfilerTest, MAYBE_WillRestartSampler) { |
908 StackSamplingProfiler::TestAPI::DisableIdleShutdown(); | |
909 | |
910 SamplingParams params; | |
911 params.sampling_interval = TimeDelta::FromMilliseconds(0); | |
912 params.samples_per_burst = 1; | |
913 | |
914 std::vector<CallStackProfile> profiles; | |
915 CaptureProfiles(params, AVeryLongTimeDelta(), &profiles); | |
916 ASSERT_EQ(1u, profiles.size()); | |
917 | |
918 // Capture thread should still be running at this point. | |
919 ASSERT_TRUE(StackSamplingProfiler::TestAPI::IsSamplingThreadRunning()); | |
920 | |
921 // Initiate an "idle" shutdown. The task will be run immediately but on | |
922 // another thread so wait for it to complete. | |
923 StackSamplingProfiler::TestAPI::InitiateSamplingThreadIdleShutdown(); | |
924 while (StackSamplingProfiler::TestAPI::IsSamplingThreadRunning()) | |
925 PlatformThread::Sleep(base::TimeDelta::FromMilliseconds(100)); | |
926 | |
927 // Ensure another capture will start the sampling thread and run. | |
928 profiles.clear(); | |
929 CaptureProfiles(params, AVeryLongTimeDelta(), &profiles); | |
930 ASSERT_EQ(1u, profiles.size()); | |
931 EXPECT_TRUE(StackSamplingProfiler::TestAPI::IsSamplingThreadRunning()); | |
932 } | |
933 | |
934 // Checks that synchronized multiple sampling requests execute in parallel. | |
935 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) | |
936 #define MAYBE_ConcurrentProfiling_InSync ConcurrentProfiling_InSync | |
937 #else | |
938 #define MAYBE_ConcurrentProfiling_InSync DISABLED_ConcurrentProfiling_InSync | |
939 #endif | |
940 TEST(StackSamplingProfilerTest, MAYBE_ConcurrentProfiling_InSync) { | |
Mike Wittman
2017/03/14 18:57:34
General comments on testing this functionality, no
bcwhite
2017/03/16 15:56:25
GetOrCreateTaskRunnerForAdd:
- state==RUNNING: tes
Mike Wittman
2017/03/18 01:38:41
Relying on existing tests is OK if the test is obv
bcwhite
2017/03/20 21:50:51
I've disabled the idle shutdown in those tests. 6
Mike Wittman
2017/03/21 16:50:38
To repeat:
Relying on existing tests is OK if the
bcwhite
2017/03/22 17:48:54
This IS the dedicated test! StopAfterIdleShutdown
Mike Wittman
2017/03/23 22:18:30
This is still pretty subtle and could use some eve
bcwhite
2017/03/27 17:52:43
Done.
Mike Wittman
2017/03/28 19:32:01
Please split this out into a dedicated test. It's
bcwhite
2017/03/29 14:56:57
That would be pretty much the same as StopDuringIn
Mike Wittman
2017/03/30 16:18:38
Yes. We might as well replace StopDuringInterSampl
bcwhite
2017/03/30 18:54:50
Done.
| |
884 WithTargetThread([](PlatformThreadId target_thread_id) { | 941 WithTargetThread([](PlatformThreadId target_thread_id) { |
885 SamplingParams params[2]; | 942 SamplingParams params[2]; |
886 params[0].initial_delay = TimeDelta::FromMilliseconds(10); | 943 params[0].initial_delay = TimeDelta::FromMilliseconds(10); |
887 params[0].sampling_interval = TimeDelta::FromMilliseconds(0); | 944 params[0].sampling_interval = TimeDelta::FromMilliseconds(1); |
888 params[0].samples_per_burst = 1; | 945 params[0].samples_per_burst = 9; |
889 | 946 |
890 params[1].sampling_interval = TimeDelta::FromMilliseconds(0); | 947 params[1].initial_delay = TimeDelta::FromMilliseconds(11); |
891 params[1].samples_per_burst = 1; | 948 params[1].sampling_interval = TimeDelta::FromMilliseconds(1); |
949 params[1].samples_per_burst = 8; | |
892 | 950 |
893 CallStackProfiles profiles[2]; | 951 CallStackProfiles profiles[2]; |
894 std::vector<std::unique_ptr<WaitableEvent>> sampling_completed(2); | 952 std::vector<std::unique_ptr<WaitableEvent>> sampling_completed( |
895 std::vector<std::unique_ptr<StackSamplingProfiler>> profiler(2); | 953 arraysize(params)); |
896 for (int i = 0; i < 2; ++i) { | 954 std::vector<std::unique_ptr<StackSamplingProfiler>> profiler( |
955 arraysize(params)); | |
956 for (size_t i = 0; i < arraysize(params); ++i) { | |
897 sampling_completed[i] = | 957 sampling_completed[i] = |
898 MakeUnique<WaitableEvent>(WaitableEvent::ResetPolicy::AUTOMATIC, | 958 MakeUnique<WaitableEvent>(WaitableEvent::ResetPolicy::AUTOMATIC, |
899 WaitableEvent::InitialState::NOT_SIGNALED); | 959 WaitableEvent::InitialState::NOT_SIGNALED); |
900 const StackSamplingProfiler::CompletedCallback callback = | 960 const StackSamplingProfiler::CompletedCallback callback = |
901 Bind(&SaveProfilesAndSignalEvent, Unretained(&profiles[i]), | 961 Bind(&SaveProfilesAndSignalEvent, Unretained(&profiles[i]), |
902 Unretained(sampling_completed[i].get())); | 962 Unretained(sampling_completed[i].get())); |
903 profiler[i] = MakeUnique<StackSamplingProfiler>(target_thread_id, | 963 profiler[i] = MakeUnique<StackSamplingProfiler>(target_thread_id, |
904 params[i], callback); | 964 params[i], callback); |
905 } | 965 } |
906 | 966 |
907 profiler[0]->Start(); | 967 profiler[0]->Start(); |
908 profiler[1]->Start(); | 968 profiler[1]->Start(); |
909 | 969 |
910 std::vector<WaitableEvent*> sampling_completed_rawptrs( | |
911 sampling_completed.size()); | |
912 std::transform( | |
913 sampling_completed.begin(), sampling_completed.end(), | |
914 sampling_completed_rawptrs.begin(), | |
915 [](const std::unique_ptr<WaitableEvent>& elem) { return elem.get(); }); | |
916 // Wait for one profiler to finish. | 970 // Wait for one profiler to finish. |
917 size_t completed_profiler = | 971 size_t completed_profiler = WaitForSamplingComplete(&sampling_completed); |
918 WaitableEvent::WaitMany(sampling_completed_rawptrs.data(), 2); | |
919 EXPECT_EQ(1u, profiles[completed_profiler].size()); | 972 EXPECT_EQ(1u, profiles[completed_profiler].size()); |
920 | 973 |
921 size_t other_profiler = 1 - completed_profiler; | 974 size_t other_profiler = 1 - completed_profiler; |
922 // Give the other profiler a chance to run and observe that it hasn't. | 975 // Wait for the other profiler to finish. |
923 EXPECT_FALSE(sampling_completed[other_profiler]->TimedWait( | |
924 TimeDelta::FromMilliseconds(25))); | |
925 | |
926 // Start the other profiler again and it should run. | |
927 profiler[other_profiler]->Start(); | |
928 sampling_completed[other_profiler]->Wait(); | 976 sampling_completed[other_profiler]->Wait(); |
929 EXPECT_EQ(1u, profiles[other_profiler].size()); | 977 EXPECT_EQ(1u, profiles[other_profiler].size()); |
978 | |
979 // Ensure each got the correct number of samples. | |
980 EXPECT_EQ(9u, profiles[0][0].samples.size()); | |
981 EXPECT_EQ(8u, profiles[1][0].samples.size()); | |
930 }); | 982 }); |
931 } | 983 } |
932 | 984 |
985 // Checks that interleaved multiple sampling requests execute in parallel. | |
986 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) | |
987 #define MAYBE_ConcurrentProfiling_Interleaved ConcurrentProfiling_Interleaved | |
988 #else | |
989 #define MAYBE_ConcurrentProfiling_Interleaved \ | |
990 DISABLED_ConcurrentProfiling_Interleaved | |
991 #endif | |
992 TEST(StackSamplingProfilerTest, MAYBE_ConcurrentProfiling_Interleaved) { | |
993 WithTargetThread([](PlatformThreadId target_thread_id) { | |
994 SamplingParams params[2]; | |
995 params[0].initial_delay = TimeDelta::FromMilliseconds(1); | |
996 params[0].sampling_interval = TimeDelta::FromMilliseconds(2); | |
997 params[0].samples_per_burst = 10; | |
998 | |
999 params[1].initial_delay = TimeDelta::FromMilliseconds(2); | |
1000 params[1].sampling_interval = TimeDelta::FromMilliseconds(2); | |
1001 params[1].samples_per_burst = 10; | |
1002 | |
1003 CallStackProfiles profiles[2]; | |
1004 std::vector<std::unique_ptr<WaitableEvent>> sampling_completed( | |
1005 arraysize(params)); | |
1006 std::vector<std::unique_ptr<StackSamplingProfiler>> profiler( | |
1007 arraysize(params)); | |
1008 for (size_t i = 0; i < arraysize(params); ++i) { | |
1009 sampling_completed[i] = | |
1010 MakeUnique<WaitableEvent>(WaitableEvent::ResetPolicy::AUTOMATIC, | |
1011 WaitableEvent::InitialState::NOT_SIGNALED); | |
1012 const StackSamplingProfiler::CompletedCallback callback = | |
1013 Bind(&SaveProfilesAndSignalEvent, Unretained(&profiles[i]), | |
1014 Unretained(sampling_completed[i].get())); | |
1015 profiler[i] = MakeUnique<StackSamplingProfiler>(target_thread_id, | |
1016 params[i], callback); | |
1017 } | |
1018 | |
1019 for (size_t i = 0; i < profiler.size(); ++i) | |
1020 profiler[i]->Start(); | |
1021 | |
1022 // Wait for one profiler to finish. | |
1023 size_t completed_profiler = WaitForSamplingComplete(&sampling_completed); | |
1024 EXPECT_EQ(1u, profiles[completed_profiler].size()); | |
1025 // Stop and destroy all profilers, always in the some order. Don't crash. | |
Mike Wittman
2017/03/14 18:57:33
nit: same
bcwhite
2017/03/16 15:56:25
Done.
| |
1026 for (size_t i = 0; i < profiler.size(); ++i) | |
1027 profiler[i]->Stop(); | |
1028 for (size_t i = 0; i < profiler.size(); ++i) | |
1029 profiler[i].reset(); | |
1030 }); | |
1031 } | |
1032 | |
1033 // Checks that several mixed sampling requests execute in parallel. | |
1034 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) | |
1035 #define MAYBE_ConcurrentProfiling_Mixed ConcurrentProfiling_Mixed | |
1036 #else | |
1037 #define MAYBE_ConcurrentProfiling_Mixed DISABLED_ConcurrentProfiling_Mixed | |
1038 #endif | |
1039 TEST(StackSamplingProfilerTest, MAYBE_ConcurrentProfiling_Mixed) { | |
1040 WithTargetThread([](PlatformThreadId target_thread_id) { | |
1041 SamplingParams params[3]; | |
1042 params[0].initial_delay = TimeDelta::FromMilliseconds(8); | |
1043 params[0].sampling_interval = TimeDelta::FromMilliseconds(4); | |
1044 params[0].samples_per_burst = 10; | |
1045 | |
1046 params[1].initial_delay = TimeDelta::FromMilliseconds(9); | |
1047 params[1].sampling_interval = TimeDelta::FromMilliseconds(3); | |
1048 params[1].samples_per_burst = 10; | |
1049 | |
1050 params[2].initial_delay = TimeDelta::FromMilliseconds(10); | |
1051 params[2].sampling_interval = TimeDelta::FromMilliseconds(2); | |
1052 params[2].samples_per_burst = 10; | |
1053 | |
1054 CallStackProfiles profiles[arraysize(params)]; | |
1055 std::vector<std::unique_ptr<WaitableEvent>> sampling_completed( | |
1056 arraysize(params)); | |
1057 std::vector<std::unique_ptr<StackSamplingProfiler>> profiler( | |
1058 arraysize(params)); | |
1059 for (size_t i = 0; i < arraysize(params); ++i) { | |
1060 sampling_completed[i] = | |
1061 MakeUnique<WaitableEvent>(WaitableEvent::ResetPolicy::AUTOMATIC, | |
1062 WaitableEvent::InitialState::NOT_SIGNALED); | |
1063 const StackSamplingProfiler::CompletedCallback callback = | |
1064 Bind(&SaveProfilesAndSignalEvent, Unretained(&profiles[i]), | |
1065 Unretained(sampling_completed[i].get())); | |
1066 profiler[i] = MakeUnique<StackSamplingProfiler>(target_thread_id, | |
1067 params[i], callback); | |
1068 } | |
1069 | |
1070 for (size_t i = 0; i < profiler.size(); ++i) | |
1071 profiler[i]->Start(); | |
1072 | |
1073 // Wait for one profiler to finish. | |
1074 size_t completed_profiler = WaitForSamplingComplete(&sampling_completed); | |
1075 EXPECT_EQ(1u, profiles[completed_profiler].size()); | |
1076 // Stop and destroy all profilers, always in the some order. Don't crash. | |
Mike Wittman
2017/03/14 18:57:34
nit: same
bcwhite
2017/03/16 15:56:25
Done.
| |
1077 for (size_t i = 0; i < profiler.size(); ++i) | |
1078 profiler[i].reset(); | |
1079 }); | |
1080 } | |
1081 | |
1082 // Checks that sampling requests execute in a staggered manner. | |
1083 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) | |
1084 #define MAYBE_ConcurrentProfiling_Staggered ConcurrentProfiling_Staggered | |
1085 #else | |
1086 #define MAYBE_ConcurrentProfiling_Staggered \ | |
1087 DISABLED_ConcurrentProfiling_Staggered | |
1088 #endif | |
1089 TEST(StackSamplingProfilerTest, MAYBE_ConcurrentProfiling_Staggered) { | |
1090 WithTargetThread([](PlatformThreadId target_thread_id) { | |
1091 SamplingParams params[3]; | |
1092 params[0].initial_delay = TimeDelta::FromMilliseconds(10); | |
1093 params[0].sampling_interval = TimeDelta::FromMilliseconds(10); | |
1094 params[0].samples_per_burst = 1; | |
1095 | |
1096 params[1].initial_delay = TimeDelta::FromMilliseconds(5); | |
1097 params[1].sampling_interval = TimeDelta::FromMilliseconds(10); | |
1098 params[1].samples_per_burst = 2; | |
1099 | |
1100 params[2].initial_delay = TimeDelta::FromMilliseconds(0); | |
1101 params[2].sampling_interval = TimeDelta::FromMilliseconds(10); | |
1102 params[2].samples_per_burst = 3; | |
1103 | |
1104 CallStackProfiles profiles[arraysize(params)]; | |
1105 std::vector<std::unique_ptr<WaitableEvent>> sampling_completed( | |
1106 arraysize(params)); | |
1107 std::vector<std::unique_ptr<StackSamplingProfiler>> profiler( | |
1108 arraysize(params)); | |
1109 for (size_t i = 0; i < arraysize(params); ++i) { | |
1110 sampling_completed[i] = | |
1111 MakeUnique<WaitableEvent>(WaitableEvent::ResetPolicy::AUTOMATIC, | |
1112 WaitableEvent::InitialState::NOT_SIGNALED); | |
1113 const StackSamplingProfiler::CompletedCallback callback = | |
1114 Bind(&SaveProfilesAndSignalEvent, Unretained(&profiles[i]), | |
1115 Unretained(sampling_completed[i].get())); | |
1116 profiler[i] = MakeUnique<StackSamplingProfiler>(target_thread_id, | |
1117 params[i], callback); | |
1118 } | |
1119 | |
1120 profiler[0]->Start(); | |
1121 profiler[1]->Start(); | |
1122 sampling_completed[0]->Wait(); | |
1123 EXPECT_FALSE(sampling_completed[1]->IsSignaled()); | |
1124 profiler[2]->Start(); | |
1125 profiler[0]->Stop(); | |
1126 profiler[1]->Stop(); | |
1127 sampling_completed[1]->Wait(); | |
1128 EXPECT_FALSE(sampling_completed[2]->IsSignaled()); | |
1129 sampling_completed[2]->Wait(); | |
1130 EXPECT_EQ(1u, profiles[0].size()); | |
1131 EXPECT_EQ(1u, profiles[1].size()); | |
1132 EXPECT_EQ(1u, profiles[2].size()); | |
1133 }); | |
1134 } | |
1135 | |
933 // Checks that a stack that runs through another library produces a stack with | 1136 // Checks that a stack that runs through another library produces a stack with |
934 // the expected functions. | 1137 // the expected functions. |
935 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) | 1138 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) |
936 #define MAYBE_OtherLibrary OtherLibrary | 1139 #define MAYBE_OtherLibrary OtherLibrary |
937 #else | 1140 #else |
938 #define MAYBE_OtherLibrary DISABLED_OtherLibrary | 1141 #define MAYBE_OtherLibrary DISABLED_OtherLibrary |
939 #endif | 1142 #endif |
940 TEST(StackSamplingProfilerTest, MAYBE_OtherLibrary) { | 1143 TEST(StackSamplingProfilerTest, MAYBE_OtherLibrary) { |
941 SamplingParams params; | 1144 SamplingParams params; |
942 params.sampling_interval = TimeDelta::FromMilliseconds(0); | 1145 params.sampling_interval = TimeDelta::FromMilliseconds(0); |
(...skipping 74 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
1017 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) | 1220 #if defined(STACK_SAMPLING_PROFILER_SUPPORTED) |
1018 #define MAYBE_UnloadedLibrary UnloadedLibrary | 1221 #define MAYBE_UnloadedLibrary UnloadedLibrary |
1019 #else | 1222 #else |
1020 #define MAYBE_UnloadedLibrary DISABLED_UnloadedLibrary | 1223 #define MAYBE_UnloadedLibrary DISABLED_UnloadedLibrary |
1021 #endif | 1224 #endif |
1022 TEST(StackSamplingProfilerTest, MAYBE_UnloadedLibrary) { | 1225 TEST(StackSamplingProfilerTest, MAYBE_UnloadedLibrary) { |
1023 TestLibraryUnload(true); | 1226 TestLibraryUnload(true); |
1024 } | 1227 } |
1025 | 1228 |
1026 } // namespace base | 1229 } // namespace base |
OLD | NEW |