Add a E2E based performance test case (#61509)
This commit is contained in:
parent
8e58c51d1a
commit
056e455e94
@ -33,7 +33,7 @@ More detailed logs should be in `build/cubic_bezier_perf.timeline.json`.
|
|||||||
To run the backdrop filter benchmark on a device:
|
To run the backdrop filter benchmark on a device:
|
||||||
To run a mobile benchmark on a device:
|
To run a mobile benchmark on a device:
|
||||||
|
|
||||||
```
|
```bash
|
||||||
flutter drive --profile -t test_driver/run_app.dart --driver test_driver/[test_name]_test.dart
|
flutter drive --profile -t test_driver/run_app.dart --driver test_driver/[test_name]_test.dart
|
||||||
```
|
```
|
||||||
|
|
||||||
@ -55,6 +55,21 @@ The key `[test_name]` can be:
|
|||||||
- `textfield_perf`
|
- `textfield_perf`
|
||||||
- `cubic_bezier_perf`
|
- `cubic_bezier_perf`
|
||||||
|
|
||||||
|
### E2E benchmarks
|
||||||
|
|
||||||
|
(On-going work)
|
||||||
|
|
||||||
|
[E2E](https://pub.dev/packages/e2e)-based tests are driven independent of the
|
||||||
|
host machine. The following tests are E2E:
|
||||||
|
|
||||||
|
- `cull_opacity_perf.dart`
|
||||||
|
|
||||||
|
These tests should be run by:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
flutter drive --profile -t test/[test_name]_e2e.dart --driver test_driver/e2e_test.dart
|
||||||
|
```
|
||||||
|
|
||||||
## Web benchmarks
|
## Web benchmarks
|
||||||
|
|
||||||
Web benchmarks are compiled from the same entry point in `lib/web_benchmarks.dart`.
|
Web benchmarks are compiled from the same entry point in `lib/web_benchmarks.dart`.
|
||||||
@ -121,3 +136,10 @@ cd dev/devicelab
|
|||||||
# Runs using the CanvasKit renderer
|
# Runs using the CanvasKit renderer
|
||||||
../../bin/cache/dart-sdk/bin/dart bin/run.dart -t bin/tasks/web_benchmarks_canvaskit.dart
|
../../bin/cache/dart-sdk/bin/dart bin/run.dart -t bin/tasks/web_benchmarks_canvaskit.dart
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## Frame policy test
|
||||||
|
|
||||||
|
File `test/frame_policy.dart` and its driving script `test_driver/frame_policy_test.dart`
|
||||||
|
are used for testing [`fullyLive`](https://api.flutter.dev/flutter/flutter_test/LiveTestWidgetsFlutterBindingFramePolicy-class.html)
|
||||||
|
and [`benchmarkLive`](https://api.flutter.dev/flutter/flutter_test/LiveTestWidgetsFlutterBindingFramePolicy-class.html)
|
||||||
|
policies in terms of its effect on [`WidgetTester.handlePointerEventRecord`](https://master-api.flutter.dev/flutter/flutter_test/WidgetTester/handlePointerEventRecord.html).
|
||||||
|
@ -0,0 +1,20 @@
|
|||||||
|
// Copyright 2014 The Flutter Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style license that can be
|
||||||
|
// found in the LICENSE file.
|
||||||
|
|
||||||
|
// The test should be run as:
|
||||||
|
// flutter drive -t test/cull_opacity_perf_e2e.dart --driver test_driver/e2e_test.dart --trace-startup --profile
|
||||||
|
|
||||||
|
import 'package:macrobenchmarks/common.dart';
|
||||||
|
|
||||||
|
import 'util.dart';
|
||||||
|
|
||||||
|
Future<void> main() async {
|
||||||
|
macroPerfTestE2E(
|
||||||
|
'cull_opacity_perf',
|
||||||
|
kCullOpacityRouteName,
|
||||||
|
pageDelay: const Duration(seconds: 1),
|
||||||
|
duration: const Duration(seconds: 10),
|
||||||
|
timeout: const Duration(seconds: 45),
|
||||||
|
);
|
||||||
|
}
|
220
dev/benchmarks/macrobenchmarks/test/util.dart
Normal file
220
dev/benchmarks/macrobenchmarks/test/util.dart
Normal file
@ -0,0 +1,220 @@
|
|||||||
|
// Copyright 2014 The Flutter Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style license that can be
|
||||||
|
// found in the LICENSE file.
|
||||||
|
|
||||||
|
import 'dart:async';
|
||||||
|
import 'dart:ui';
|
||||||
|
|
||||||
|
import 'package:flutter/scheduler.dart';
|
||||||
|
import 'package:flutter_test/flutter_test.dart';
|
||||||
|
import 'package:flutter/widgets.dart';
|
||||||
|
|
||||||
|
import 'package:macrobenchmarks/common.dart';
|
||||||
|
import 'package:e2e/e2e.dart';
|
||||||
|
import 'package:macrobenchmarks/main.dart' as app;
|
||||||
|
|
||||||
|
/// The maximum amount of time considered safe to spend for a frame's build
|
||||||
|
/// phase. Anything past that is in the danger of missing the frame as 60FPS.
|
||||||
|
///
|
||||||
|
/// Changing this doesn't re-evaluate existing summary.
|
||||||
|
Duration kBuildBudget = const Duration(milliseconds: 16);
|
||||||
|
// TODO(CareF): Automatically calculate the refresh budget (#61958)
|
||||||
|
|
||||||
|
typedef ControlCallback = Future<void> Function(WidgetController controller);
|
||||||
|
|
||||||
|
void macroPerfTestE2E(
|
||||||
|
String testName,
|
||||||
|
String routeName, {
|
||||||
|
Duration pageDelay,
|
||||||
|
Duration duration = const Duration(seconds: 3),
|
||||||
|
Duration timeout = const Duration(seconds: 30),
|
||||||
|
ControlCallback body,
|
||||||
|
ControlCallback setup,
|
||||||
|
}) {
|
||||||
|
assert(() {
|
||||||
|
debugPrint(kDebugWarning);
|
||||||
|
return true;
|
||||||
|
}());
|
||||||
|
final WidgetsBinding _binding = E2EWidgetsFlutterBinding.ensureInitialized();
|
||||||
|
assert(_binding is E2EWidgetsFlutterBinding);
|
||||||
|
final E2EWidgetsFlutterBinding binding = _binding as E2EWidgetsFlutterBinding;
|
||||||
|
binding.framePolicy = LiveTestWidgetsFlutterBindingFramePolicy.benchmarkLive;
|
||||||
|
|
||||||
|
testWidgets(testName, (WidgetTester tester) async {
|
||||||
|
assert((tester.binding as LiveTestWidgetsFlutterBinding).framePolicy ==
|
||||||
|
LiveTestWidgetsFlutterBindingFramePolicy.fullyLive);
|
||||||
|
app.main();
|
||||||
|
await tester.pumpAndSettle();
|
||||||
|
|
||||||
|
// The slight initial delay avoids starting the timing during a
|
||||||
|
// period of increased load on the device. Without this delay, the
|
||||||
|
// benchmark has greater noise.
|
||||||
|
// See: https://github.com/flutter/flutter/issues/19434
|
||||||
|
await tester.binding.delayed(const Duration(microseconds: 250));
|
||||||
|
|
||||||
|
final Finder scrollable =
|
||||||
|
find.byKey(const ValueKey<String>(kScrollableName));
|
||||||
|
expect(scrollable, findsOneWidget);
|
||||||
|
final Finder button =
|
||||||
|
find.byKey(ValueKey<String>(routeName), skipOffstage: false);
|
||||||
|
await tester.ensureVisible(button);
|
||||||
|
expect(button, findsOneWidget);
|
||||||
|
await tester.pumpAndSettle();
|
||||||
|
await tester.tap(button);
|
||||||
|
|
||||||
|
if (pageDelay != null) {
|
||||||
|
// Wait for the page to load
|
||||||
|
await tester.binding.delayed(pageDelay);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (setup != null) {
|
||||||
|
await setup(tester);
|
||||||
|
}
|
||||||
|
|
||||||
|
await watchPerformance(binding, () async {
|
||||||
|
final Future<void> durationFuture = tester.binding.delayed(duration);
|
||||||
|
if (body != null) {
|
||||||
|
await body(tester);
|
||||||
|
}
|
||||||
|
await durationFuture;
|
||||||
|
});
|
||||||
|
}, semanticsEnabled: false, timeout: Timeout(timeout));
|
||||||
|
}
|
||||||
|
|
||||||
|
Future<void> watchPerformance(
|
||||||
|
E2EWidgetsFlutterBinding binding,
|
||||||
|
Future<void> action(),
|
||||||
|
) async {
|
||||||
|
final List<FrameTiming> frameTimings = <FrameTiming>[];
|
||||||
|
final TimingsCallback watcher = frameTimings.addAll;
|
||||||
|
binding.addTimingsCallback(watcher);
|
||||||
|
await action();
|
||||||
|
binding.removeTimingsCallback(watcher);
|
||||||
|
// TODO(CareF): determine if it's running on firebase and report metric online
|
||||||
|
final FrameTimingSummarizer frameTimes = FrameTimingSummarizer(frameTimings);
|
||||||
|
binding.reportData = <String, dynamic>{'performance': frameTimes.summary};
|
||||||
|
}
|
||||||
|
|
||||||
|
/// This class and summarizes a list of [FrameTiming] for the performance
|
||||||
|
/// metrics.
|
||||||
|
class FrameTimingSummarizer {
|
||||||
|
factory FrameTimingSummarizer(List<FrameTiming> data) {
|
||||||
|
assert(data != null);
|
||||||
|
assert(data.isNotEmpty);
|
||||||
|
final List<Duration> frameBuildTime = List<Duration>.unmodifiable(
|
||||||
|
data.map<Duration>((FrameTiming datum) => datum.buildDuration),
|
||||||
|
);
|
||||||
|
final List<Duration> frameBuildTimeSorted = List<Duration>.from(frameBuildTime)..sort();
|
||||||
|
final List<Duration> frameRasterizerTime = List<Duration>.unmodifiable(
|
||||||
|
data.map<Duration>((FrameTiming datum) => datum.rasterDuration),
|
||||||
|
);
|
||||||
|
final List<Duration> frameRasterizerTimeSorted = List<Duration>.from(frameBuildTime)..sort();
|
||||||
|
final Duration Function(Duration, Duration) add = (Duration a, Duration b) => a + b;
|
||||||
|
return FrameTimingSummarizer._(
|
||||||
|
frameBuildTime: frameBuildTime,
|
||||||
|
frameRasterizerTime: frameRasterizerTime,
|
||||||
|
// This avarage calculation is microsecond precision, which is fine
|
||||||
|
// because typical values of these times are milliseconds.
|
||||||
|
averageFrameBuildTime: frameBuildTime.reduce(add) ~/ data.length,
|
||||||
|
p90FrameBuildTime: _findPercentile(frameBuildTimeSorted, 0.90),
|
||||||
|
p99FrameBuildTime: _findPercentile(frameBuildTimeSorted, 0.99),
|
||||||
|
worstFrameBuildTime: frameBuildTimeSorted.last,
|
||||||
|
missedFrameBuildBudget: _countExceed(frameBuildTimeSorted, kBuildBudget),
|
||||||
|
averageFrameRasterizerTime: frameRasterizerTime.reduce(add) ~/ data.length,
|
||||||
|
p90FrameRasterizerTime: _findPercentile(frameRasterizerTimeSorted, 0.90),
|
||||||
|
p99FrameRasterizerTime: _findPercentile(frameRasterizerTimeSorted, 0.90),
|
||||||
|
worstFrameRasterizerTime: frameRasterizerTimeSorted.last,
|
||||||
|
missedFrameRasterizerBudget: _countExceed(frameRasterizerTimeSorted, kBuildBudget),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
const FrameTimingSummarizer._({
|
||||||
|
@required this.frameBuildTime,
|
||||||
|
@required this.frameRasterizerTime,
|
||||||
|
@required this.averageFrameBuildTime,
|
||||||
|
@required this.p90FrameBuildTime,
|
||||||
|
@required this.p99FrameBuildTime,
|
||||||
|
@required this.worstFrameBuildTime,
|
||||||
|
@required this.missedFrameBuildBudget,
|
||||||
|
@required this.averageFrameRasterizerTime,
|
||||||
|
@required this.p90FrameRasterizerTime,
|
||||||
|
@required this.p99FrameRasterizerTime,
|
||||||
|
@required this.worstFrameRasterizerTime,
|
||||||
|
@required this.missedFrameRasterizerBudget
|
||||||
|
});
|
||||||
|
|
||||||
|
/// List of frame build time in microseconds
|
||||||
|
final List<Duration> frameBuildTime;
|
||||||
|
|
||||||
|
/// List of frame rasterizer time in microseconds
|
||||||
|
final List<Duration> frameRasterizerTime;
|
||||||
|
|
||||||
|
/// The average value of [frameBuildTime] in milliseconds.
|
||||||
|
final Duration averageFrameBuildTime;
|
||||||
|
|
||||||
|
/// The 90-th percentile value of [frameBuildTime] in milliseconds
|
||||||
|
final Duration p90FrameBuildTime;
|
||||||
|
|
||||||
|
/// The 99-th percentile value of [frameBuildTime] in milliseconds
|
||||||
|
final Duration p99FrameBuildTime;
|
||||||
|
|
||||||
|
/// The largest value of [frameBuildTime] in milliseconds
|
||||||
|
final Duration worstFrameBuildTime;
|
||||||
|
|
||||||
|
/// Number of items in [frameBuildTime] that's greater than [kBuildBudget]
|
||||||
|
final int missedFrameBuildBudget;
|
||||||
|
|
||||||
|
/// The average value of [frameRasterizerTime] in milliseconds.
|
||||||
|
final Duration averageFrameRasterizerTime;
|
||||||
|
|
||||||
|
/// The 90-th percentile value of [frameRasterizerTime] in milliseconds.
|
||||||
|
final Duration p90FrameRasterizerTime;
|
||||||
|
|
||||||
|
/// The 99-th percentile value of [frameRasterizerTime] in milliseconds.
|
||||||
|
final Duration p99FrameRasterizerTime;
|
||||||
|
|
||||||
|
/// The largest value of [frameRasterizerTime] in milliseconds.
|
||||||
|
final Duration worstFrameRasterizerTime;
|
||||||
|
|
||||||
|
/// Number of items in [frameRasterizerTime] that's greater than [kBuildBudget]
|
||||||
|
final int missedFrameRasterizerBudget;
|
||||||
|
|
||||||
|
Map<String, dynamic> get summary => <String, dynamic>{
|
||||||
|
'average_frame_build_time_millis':
|
||||||
|
averageFrameBuildTime.inMicroseconds / 1E3,
|
||||||
|
'90th_percentile_frame_build_time_millis':
|
||||||
|
p90FrameBuildTime.inMicroseconds / 1E3,
|
||||||
|
'99th_percentile_frame_build_time_millis':
|
||||||
|
p99FrameBuildTime.inMicroseconds / 1E3,
|
||||||
|
'worst_frame_build_time_millis':
|
||||||
|
worstFrameBuildTime.inMicroseconds / 1E3,
|
||||||
|
'missed_frame_build_budget_count': missedFrameBuildBudget,
|
||||||
|
'average_frame_rasterizer_time_millis':
|
||||||
|
averageFrameRasterizerTime.inMicroseconds / 1E3,
|
||||||
|
'90th_percentile_frame_rasterizer_time_millis':
|
||||||
|
p90FrameRasterizerTime.inMicroseconds / 1E3,
|
||||||
|
'99th_percentile_frame_rasterizer_time_millis':
|
||||||
|
p99FrameRasterizerTime.inMicroseconds / 1E3,
|
||||||
|
'worst_frame_rasterizer_time_millis':
|
||||||
|
worstFrameRasterizerTime.inMicroseconds / 1E3,
|
||||||
|
'missed_frame_rasterizer_budget_count': missedFrameRasterizerBudget,
|
||||||
|
'frame_count': frameBuildTime.length,
|
||||||
|
'frame_build_times': frameBuildTime
|
||||||
|
.map<int>((Duration datum) => datum.inMicroseconds).toList(),
|
||||||
|
'frame_rasterizer_times': frameRasterizerTime
|
||||||
|
.map<int>((Duration datum) => datum.inMicroseconds).toList(),
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
// The following helper functions require data sorted
|
||||||
|
|
||||||
|
// return the 100*p-th percentile of the data
|
||||||
|
T _findPercentile<T>(List<T> data, double p) {
|
||||||
|
assert(p >= 0 && p <= 1);
|
||||||
|
return data[((data.length - 1) * p).round()];
|
||||||
|
}
|
||||||
|
|
||||||
|
// return the number of items in data that > threshold
|
||||||
|
int _countExceed<T extends Comparable<T>>(List<T> data, T threshold) {
|
||||||
|
return data.length - data.indexWhere((T datum) => datum.compareTo(threshold) > 0);
|
||||||
|
}
|
57
dev/benchmarks/macrobenchmarks/test_driver/e2e_test.dart
Normal file
57
dev/benchmarks/macrobenchmarks/test_driver/e2e_test.dart
Normal file
@ -0,0 +1,57 @@
|
|||||||
|
// Copyright 2014 The Flutter Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style license that can be
|
||||||
|
// found in the LICENSE file.
|
||||||
|
|
||||||
|
import 'dart:async';
|
||||||
|
import 'dart:convert';
|
||||||
|
import 'dart:io';
|
||||||
|
|
||||||
|
import 'package:e2e/common.dart' as e2e;
|
||||||
|
import 'package:flutter_driver/flutter_driver.dart';
|
||||||
|
|
||||||
|
import 'package:path/path.dart' as path;
|
||||||
|
|
||||||
|
const JsonEncoder _prettyEncoder = JsonEncoder.withIndent(' ');
|
||||||
|
|
||||||
|
/// Flutter Driver test output directory.
|
||||||
|
///
|
||||||
|
/// Tests should write any output files to this directory. Defaults to the path
|
||||||
|
/// set in the FLUTTER_TEST_OUTPUTS_DIR environment variable, or `build` if
|
||||||
|
/// unset.
|
||||||
|
String testOutputsDirectory = Platform.environment['FLUTTER_TEST_OUTPUTS_DIR'] ?? 'build';
|
||||||
|
|
||||||
|
String testOutputFilename = 'e2e_perf_summary';
|
||||||
|
|
||||||
|
Future<void> main() async {
|
||||||
|
final FlutterDriver driver = await FlutterDriver.connect();
|
||||||
|
final String jsonResult =
|
||||||
|
await driver.requestData(null, timeout: const Duration(minutes: 1));
|
||||||
|
final e2e.Response response = e2e.Response.fromJson(jsonResult);
|
||||||
|
await driver.close();
|
||||||
|
|
||||||
|
if (response.allTestsPassed) {
|
||||||
|
print('All tests passed.');
|
||||||
|
|
||||||
|
await fs.directory(testOutputsDirectory).create(recursive: true);
|
||||||
|
final File file = fs.file(path.join(
|
||||||
|
testOutputsDirectory,
|
||||||
|
'$testOutputFilename.json'
|
||||||
|
));
|
||||||
|
final String resultString = _encodeJson(
|
||||||
|
response.data['performance'] as Map<String, dynamic>,
|
||||||
|
true,
|
||||||
|
);
|
||||||
|
await file.writeAsString(resultString);
|
||||||
|
|
||||||
|
exit(0);
|
||||||
|
} else {
|
||||||
|
print('Failure Details:\n${response.formattedFailureDetails}');
|
||||||
|
exit(1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
String _encodeJson(Map<String, dynamic> jsonObject, bool pretty) {
|
||||||
|
return pretty
|
||||||
|
? _prettyEncoder.convert(jsonObject)
|
||||||
|
: json.encode(jsonObject);
|
||||||
|
}
|
14
dev/devicelab/bin/tasks/cull_opacity_perf__e2e_summary.dart
Normal file
14
dev/devicelab/bin/tasks/cull_opacity_perf__e2e_summary.dart
Normal file
@ -0,0 +1,14 @@
|
|||||||
|
// Copyright 2014 The Flutter Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style license that can be
|
||||||
|
// found in the LICENSE file.
|
||||||
|
|
||||||
|
import 'dart:async';
|
||||||
|
|
||||||
|
import 'package:flutter_devicelab/tasks/perf_tests.dart';
|
||||||
|
import 'package:flutter_devicelab/framework/adb.dart';
|
||||||
|
import 'package:flutter_devicelab/framework/framework.dart';
|
||||||
|
|
||||||
|
Future<void> main() async {
|
||||||
|
deviceOperatingSystem = DeviceOperatingSystem.android;
|
||||||
|
await task(createCullOpacityPerfE2ETest());
|
||||||
|
}
|
@ -76,6 +76,13 @@ TaskFunction createCullOpacityPerfTest() {
|
|||||||
).run;
|
).run;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
TaskFunction createCullOpacityPerfE2ETest() {
|
||||||
|
return E2EPerfTest(
|
||||||
|
'${flutterDirectory.path}/dev/benchmarks/macrobenchmarks',
|
||||||
|
'test/cull_opacity_perf_e2e.dart',
|
||||||
|
).run;
|
||||||
|
}
|
||||||
|
|
||||||
TaskFunction createCubicBezierPerfTest() {
|
TaskFunction createCubicBezierPerfTest() {
|
||||||
return PerfTest(
|
return PerfTest(
|
||||||
'${flutterDirectory.path}/dev/benchmarks/macrobenchmarks',
|
'${flutterDirectory.path}/dev/benchmarks/macrobenchmarks',
|
||||||
@ -348,6 +355,8 @@ class PerfTest {
|
|||||||
this.timelineFileName, {
|
this.timelineFileName, {
|
||||||
this.needsMeasureCpuGpu = false,
|
this.needsMeasureCpuGpu = false,
|
||||||
this.testDriver,
|
this.testDriver,
|
||||||
|
this.needsFullTimeline = true,
|
||||||
|
this.benchmarkScoreKeys,
|
||||||
});
|
});
|
||||||
|
|
||||||
/// The directory where the app under test is defined.
|
/// The directory where the app under test is defined.
|
||||||
@ -356,10 +365,35 @@ class PerfTest {
|
|||||||
final String testTarget;
|
final String testTarget;
|
||||||
// The prefix name of the filename such as `<timelineFileName>.timeline_summary.json`.
|
// The prefix name of the filename such as `<timelineFileName>.timeline_summary.json`.
|
||||||
final String timelineFileName;
|
final String timelineFileName;
|
||||||
|
String get resultFilename => '$timelineFileName.timeline_summary';
|
||||||
/// The test file to run on the host.
|
/// The test file to run on the host.
|
||||||
final String testDriver;
|
final String testDriver;
|
||||||
/// Whether to collect CPU and GPU metrics.
|
/// Whether to collect CPU and GPU metrics.
|
||||||
final bool needsMeasureCpuGpu;
|
final bool needsMeasureCpuGpu;
|
||||||
|
/// Whether to collect full timeline, meaning if `--trace-startup` flag is needed.
|
||||||
|
final bool needsFullTimeline;
|
||||||
|
|
||||||
|
/// The keys of the values that need to be reported.
|
||||||
|
///
|
||||||
|
/// If it's `null`, then report:
|
||||||
|
/// ```Dart
|
||||||
|
/// <String>[
|
||||||
|
/// 'average_frame_build_time_millis',
|
||||||
|
/// 'worst_frame_build_time_millis',
|
||||||
|
/// '90th_percentile_frame_build_time_millis',
|
||||||
|
/// '99th_percentile_frame_build_time_millis',
|
||||||
|
/// 'average_frame_rasterizer_time_millis',
|
||||||
|
/// 'worst_frame_rasterizer_time_millis',
|
||||||
|
/// '90th_percentile_frame_rasterizer_time_millis',
|
||||||
|
/// '99th_percentile_frame_rasterizer_time_millis',
|
||||||
|
/// 'average_vsync_transitions_missed',
|
||||||
|
/// '90th_percentile_vsync_transitions_missed',
|
||||||
|
/// '99th_percentile_vsync_transitions_missed',
|
||||||
|
/// if (needsMeasureCpuGpu) 'cpu_percentage',
|
||||||
|
/// if (needsMeasureCpuGpu) 'gpu_percentage',
|
||||||
|
/// ]
|
||||||
|
/// ```
|
||||||
|
final List<String> benchmarkScoreKeys;
|
||||||
|
|
||||||
Future<TaskResult> run() {
|
Future<TaskResult> run() {
|
||||||
return internalRun();
|
return internalRun();
|
||||||
@ -382,7 +416,8 @@ class PerfTest {
|
|||||||
'-v',
|
'-v',
|
||||||
'--verbose-system-logs',
|
'--verbose-system-logs',
|
||||||
'--profile',
|
'--profile',
|
||||||
'--trace-startup', // Enables "endless" timeline event buffering.
|
if (needsFullTimeline)
|
||||||
|
'--trace-startup', // Enables "endless" timeline event buffering.
|
||||||
'-t', testTarget,
|
'-t', testTarget,
|
||||||
if (noBuild) '--no-build',
|
if (noBuild) '--no-build',
|
||||||
if (testDriver != null)
|
if (testDriver != null)
|
||||||
@ -396,7 +431,7 @@ class PerfTest {
|
|||||||
deviceId,
|
deviceId,
|
||||||
]);
|
]);
|
||||||
final Map<String, dynamic> data = json.decode(
|
final Map<String, dynamic> data = json.decode(
|
||||||
file('$testDirectory/build/$timelineFileName.timeline_summary.json').readAsStringSync(),
|
file('$testDirectory/build/$resultFilename.json').readAsStringSync(),
|
||||||
) as Map<String, dynamic>;
|
) as Map<String, dynamic>;
|
||||||
|
|
||||||
if (data['frame_count'] as int < 5) {
|
if (data['frame_count'] as int < 5) {
|
||||||
@ -412,25 +447,57 @@ class PerfTest {
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
return TaskResult.success(data, benchmarkScoreKeys: <String>[
|
return TaskResult.success(
|
||||||
'average_frame_build_time_millis',
|
data,
|
||||||
'worst_frame_build_time_millis',
|
benchmarkScoreKeys: benchmarkScoreKeys ?? <String>[
|
||||||
'90th_percentile_frame_build_time_millis',
|
'average_frame_build_time_millis',
|
||||||
'99th_percentile_frame_build_time_millis',
|
'worst_frame_build_time_millis',
|
||||||
'average_frame_rasterizer_time_millis',
|
'90th_percentile_frame_build_time_millis',
|
||||||
'worst_frame_rasterizer_time_millis',
|
'99th_percentile_frame_build_time_millis',
|
||||||
'90th_percentile_frame_rasterizer_time_millis',
|
'average_frame_rasterizer_time_millis',
|
||||||
'99th_percentile_frame_rasterizer_time_millis',
|
'worst_frame_rasterizer_time_millis',
|
||||||
'average_vsync_transitions_missed',
|
'90th_percentile_frame_rasterizer_time_millis',
|
||||||
'90th_percentile_vsync_transitions_missed',
|
'99th_percentile_frame_rasterizer_time_millis',
|
||||||
'99th_percentile_vsync_transitions_missed',
|
'average_vsync_transitions_missed',
|
||||||
if (needsMeasureCpuGpu) 'cpu_percentage',
|
'90th_percentile_vsync_transitions_missed',
|
||||||
if (needsMeasureCpuGpu) 'gpu_percentage',
|
'99th_percentile_vsync_transitions_missed',
|
||||||
]);
|
if (needsMeasureCpuGpu) 'cpu_percentage',
|
||||||
|
if (needsMeasureCpuGpu) 'gpu_percentage',
|
||||||
|
],
|
||||||
|
);
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
class E2EPerfTest extends PerfTest {
|
||||||
|
const E2EPerfTest(
|
||||||
|
String testDirectory,
|
||||||
|
String testTarget, {
|
||||||
|
String summaryFilename,
|
||||||
|
List<String> benchmarkScoreKeys,
|
||||||
|
}
|
||||||
|
) : super(
|
||||||
|
testDirectory,
|
||||||
|
testTarget,
|
||||||
|
summaryFilename,
|
||||||
|
testDriver: 'test_driver/e2e_test.dart',
|
||||||
|
needsFullTimeline: false,
|
||||||
|
benchmarkScoreKeys: benchmarkScoreKeys ?? const <String>[
|
||||||
|
'average_frame_build_time_millis',
|
||||||
|
'worst_frame_build_time_millis',
|
||||||
|
'90th_percentile_frame_build_time_millis',
|
||||||
|
'99th_percentile_frame_build_time_millis',
|
||||||
|
'average_frame_rasterizer_time_millis',
|
||||||
|
'worst_frame_rasterizer_time_millis',
|
||||||
|
'90th_percentile_frame_rasterizer_time_millis',
|
||||||
|
'99th_percentile_frame_rasterizer_time_millis',
|
||||||
|
],
|
||||||
|
);
|
||||||
|
|
||||||
|
@override
|
||||||
|
String get resultFilename => timelineFileName ?? 'e2e_perf_summary';
|
||||||
|
}
|
||||||
|
|
||||||
class PerfTestWithSkSL extends PerfTest {
|
class PerfTestWithSkSL extends PerfTest {
|
||||||
PerfTestWithSkSL(
|
PerfTestWithSkSL(
|
||||||
String testDirectory,
|
String testDirectory,
|
||||||
|
@ -161,6 +161,13 @@ tasks:
|
|||||||
stage: devicelab
|
stage: devicelab
|
||||||
required_agent_capabilities: ["mac/android"]
|
required_agent_capabilities: ["mac/android"]
|
||||||
|
|
||||||
|
cull_opacity_perf__e2e_summary:
|
||||||
|
description: >
|
||||||
|
Measures the runtime performance of culling opacity widgets on Android on
|
||||||
|
E2E with self-driving test app.
|
||||||
|
stage: devicelab
|
||||||
|
required_agent_capabilities: ["linux/android"]
|
||||||
|
|
||||||
multi_widget_construction_perf__timeline_summary:
|
multi_widget_construction_perf__timeline_summary:
|
||||||
description: >
|
description: >
|
||||||
Measures the runtime performance of constructing and destructing widgets on Android.
|
Measures the runtime performance of constructing and destructing widgets on Android.
|
||||||
|
@ -270,6 +270,22 @@ class TargetPlatformVariant extends TestVariant<TargetPlatform> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// The warning message to show when a benchmark is performed with assert on.
|
||||||
|
const String kDebugWarning = '''
|
||||||
|
┏╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍┓
|
||||||
|
┇ ⚠ THIS BENCHMARK IS BEING RUN IN DEBUG MODE ⚠ ┇
|
||||||
|
┡╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍┦
|
||||||
|
│ │
|
||||||
|
│ Numbers obtained from a benchmark while asserts are │
|
||||||
|
│ enabled will not accurately reflect the performance │
|
||||||
|
│ that will be experienced by end users using release ╎
|
||||||
|
│ builds. Benchmarks should be run using this command ╎
|
||||||
|
│ line: "flutter run --profile test.dart" or ┊
|
||||||
|
│ or "flutter drive --profile -t test.dart". ┊
|
||||||
|
│ ┊
|
||||||
|
└─────────────────────────────────────────────────╌┄┈ 🐢
|
||||||
|
''';
|
||||||
|
|
||||||
/// Runs the [callback] inside the Flutter benchmark environment.
|
/// Runs the [callback] inside the Flutter benchmark environment.
|
||||||
///
|
///
|
||||||
/// Use this function for benchmarking custom [StatelessWidget]s and
|
/// Use this function for benchmarking custom [StatelessWidget]s and
|
||||||
@ -320,18 +336,7 @@ Future<void> benchmarkWidgets(
|
|||||||
assert(() {
|
assert(() {
|
||||||
if (mayRunWithAsserts)
|
if (mayRunWithAsserts)
|
||||||
return true;
|
return true;
|
||||||
|
print(kDebugWarning);
|
||||||
print('┏╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍┓');
|
|
||||||
print('┇ ⚠ THIS BENCHMARK IS BEING RUN WITH ASSERTS ENABLED ⚠ ┇');
|
|
||||||
print('┡╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍╍┦');
|
|
||||||
print('│ │');
|
|
||||||
print('│ Numbers obtained from a benchmark while asserts are │');
|
|
||||||
print('│ enabled will not accurately reflect the performance │');
|
|
||||||
print('│ that will be experienced by end users using release ╎');
|
|
||||||
print('│ builds. Benchmarks should be run using this command ┆');
|
|
||||||
print('│ line: flutter run --release benchmark.dart ┊');
|
|
||||||
print('│ ');
|
|
||||||
print('└─────────────────────────────────────────────────╌┄┈ 🐢');
|
|
||||||
return true;
|
return true;
|
||||||
}());
|
}());
|
||||||
final TestWidgetsFlutterBinding binding = TestWidgetsFlutterBinding.ensureInitialized() as TestWidgetsFlutterBinding;
|
final TestWidgetsFlutterBinding binding = TestWidgetsFlutterBinding.ensureInitialized() as TestWidgetsFlutterBinding;
|
||||||
|
Loading…
x
Reference in New Issue
Block a user