Merge "docs: format several docs" into main
diff --git a/protos/third_party/chromium/chrome_track_event.proto b/protos/third_party/chromium/chrome_track_event.proto
index 7cf29f6..6d92cb6 100644
--- a/protos/third_party/chromium/chrome_track_event.proto
+++ b/protos/third_party/chromium/chrome_track_event.proto
@@ -2020,6 +2020,15 @@
 
   // Timestamp in microseconds of the start of the task containing this slice.
   optional uint64 task_start_time_us = 2;
+
+  // t1 - t0, where t1 is the start timestamp of this slice and t0 is the
+  // timestamp of the time when the task containing this slice
+  // was queued.
+  optional uint64 task_queueing_time_us = 3;
+
+  // Timestamp in microseconds of the time when the task containing
+  // this slice was queued.
+  optional uint64 task_queued_time_us = 4;
 }
 
 message ChromeLatencyInfo2 {
diff --git a/ui/src/frontend/base_slice_track.ts b/ui/src/frontend/base_slice_track.ts
index 45cbc9e..0c18970 100644
--- a/ui/src/frontend/base_slice_track.ts
+++ b/ui/src/frontend/base_slice_track.ts
@@ -35,7 +35,7 @@
 import {TrackMouseEvent, TrackRenderContext} from '../public/track';
 import {Point2D, VerticalBounds} from '../base/geom';
 import {Trace} from '../public/trace';
-import {Ds} from '../trace_processor/dataset';
+import {SourceDataset, Dataset} from '../trace_processor/dataset';
 
 // The common class that underpins all tracks drawing slices.
 
@@ -974,15 +974,15 @@
     return {ts: Time.fromRaw(row.ts), dur: Duration.fromRaw(row.dur)};
   }
 
-  getDataset(): Ds.Dataset | undefined {
-    return {
+  getDataset(): Dataset | undefined {
+    return new SourceDataset({
       src: this.getSqlSource(),
       schema: {
         id: NUM,
         ts: LONG,
         dur: LONG,
       },
-    };
+    });
   }
 }
 
diff --git a/ui/src/frontend/named_slice_track.ts b/ui/src/frontend/named_slice_track.ts
index edb7ec9..7a23285 100644
--- a/ui/src/frontend/named_slice_track.ts
+++ b/ui/src/frontend/named_slice_track.ts
@@ -30,7 +30,7 @@
 import {renderDuration} from './widgets/duration';
 import {TraceImpl} from '../core/trace_impl';
 import {assertIsInstance} from '../base/logging';
-import {Ds} from '../trace_processor/dataset';
+import {SourceDataset, Dataset} from '../trace_processor/dataset';
 
 export const NAMED_ROW = {
   // Base columns (tsq, ts, dur, id, depth).
@@ -82,8 +82,8 @@
     return new ThreadSliceDetailsPanel(assertIsInstance(this.trace, TraceImpl));
   }
 
-  override getDataset(): Ds.Dataset | undefined {
-    return {
+  override getDataset(): Dataset | undefined {
+    return new SourceDataset({
       src: this.getSqlSource(),
       schema: {
         id: NUM,
@@ -91,6 +91,6 @@
         ts: LONG,
         dur: LONG,
       },
-    };
+    });
   }
 }
diff --git a/ui/src/plugins/dev.perfetto.AsyncSlices/async_slice_track.ts b/ui/src/plugins/dev.perfetto.AsyncSlices/async_slice_track.ts
index 4c8898b..0486e6e 100644
--- a/ui/src/plugins/dev.perfetto.AsyncSlices/async_slice_track.ts
+++ b/ui/src/plugins/dev.perfetto.AsyncSlices/async_slice_track.ts
@@ -14,12 +14,12 @@
 
 import {BigintMath as BIMath} from '../../base/bigint_math';
 import {clamp} from '../../base/math_utils';
-import {Ds} from '../../trace_processor/dataset';
 import {NAMED_ROW, NamedSliceTrack} from '../../frontend/named_slice_track';
 import {SLICE_LAYOUT_FIT_CONTENT_DEFAULTS} from '../../frontend/slice_layout';
 import {NewTrackArgs} from '../../frontend/track';
 import {TrackEventDetails} from '../../public/selection';
 import {Slice} from '../../public/track';
+import {SourceDataset, Dataset} from '../../trace_processor/dataset';
 import {
   LONG,
   LONG_NULL,
@@ -112,8 +112,8 @@
     };
   }
 
-  override getDataset(): Ds.Dataset {
-    return {
+  override getDataset(): Dataset {
+    return new SourceDataset({
       src: `slice`,
       filter: {
         col: 'track_id',
@@ -126,6 +126,6 @@
         dur: LONG,
         parent_id: NUM_NULL,
       },
-    };
+    });
   }
 }
diff --git a/ui/src/plugins/dev.perfetto.AsyncSlices/slice_selection_aggregator.ts b/ui/src/plugins/dev.perfetto.AsyncSlices/slice_selection_aggregator.ts
index 0364152..23226bc 100644
--- a/ui/src/plugins/dev.perfetto.AsyncSlices/slice_selection_aggregator.ts
+++ b/ui/src/plugins/dev.perfetto.AsyncSlices/slice_selection_aggregator.ts
@@ -16,7 +16,7 @@
 import {AreaSelection} from '../../public/selection';
 import {Engine} from '../../trace_processor/engine';
 import {AreaSelectionAggregator} from '../../public/selection';
-import {Ds} from '../../trace_processor/dataset';
+import {UnionDataset} from '../../trace_processor/dataset';
 import {LONG, NUM, STR} from '../../trace_processor/query_result';
 
 export class SliceSelectionAggregator implements AreaSelectionAggregator {
@@ -30,13 +30,13 @@
       dur: LONG,
     };
     const validDatasets = area.tracks
-      .map((t) => t.track.getDataset?.())
-      .filter((d) => d !== undefined)
-      .filter((d) => Ds.doesImplement(d, desiredSchema));
+      .map((track) => track.track.getDataset?.())
+      .filter((ds) => ds !== undefined)
+      .filter((ds) => ds.implements(desiredSchema));
     if (validDatasets.length === 0) {
       return false;
     }
-    const optimizedDataset = Ds.optimize({union: validDatasets});
+    const unionDataset = new UnionDataset(validDatasets);
     await engine.query(`
       create or replace perfetto table ${this.id} as
       select
@@ -44,7 +44,7 @@
         sum(dur) AS total_dur,
         sum(dur)/count() as avg_dur,
         count() as occurrences
-        from (${Ds.query(optimizedDataset)})
+        from (${unionDataset.optimize().query()})
       where
         ts + dur > ${area.start}
         and ts < ${area.end}
diff --git a/ui/src/plugins/dev.perfetto.Ftrace/ftrace_track.ts b/ui/src/plugins/dev.perfetto.Ftrace/ftrace_track.ts
index d056a8d..4ef7793 100644
--- a/ui/src/plugins/dev.perfetto.Ftrace/ftrace_track.ts
+++ b/ui/src/plugins/dev.perfetto.Ftrace/ftrace_track.ts
@@ -24,7 +24,7 @@
 import {FtraceFilter} from './common';
 import {Monitor} from '../../base/monitor';
 import {TrackRenderContext} from '../../public/track';
-import {Ds} from '../../trace_processor/dataset';
+import {SourceDataset, Dataset} from '../../trace_processor/dataset';
 
 const MARGIN = 2;
 const RECT_HEIGHT = 18;
@@ -57,8 +57,8 @@
     this.monitor = new Monitor([() => store.state]);
   }
 
-  getDataset(): Ds.Dataset {
-    return {
+  getDataset(): Dataset {
+    return new SourceDataset({
       // 'ftrace_event' doesn't have a dur column, but injecting dur=0 (all
       // ftrace events are effectively 'instant') allows us to participate in
       // generic slice aggregations
@@ -73,7 +73,7 @@
         col: 'cpu',
         eq: this.cpu,
       },
-    };
+    });
   }
 
   async onUpdate({
diff --git a/ui/src/public/track.ts b/ui/src/public/track.ts
index 93d493e..6d1b1dc 100644
--- a/ui/src/public/track.ts
+++ b/ui/src/public/track.ts
@@ -20,7 +20,7 @@
 import {ColorScheme} from './color_scheme';
 import {TrackEventDetailsPanel} from './details_panel';
 import {TrackEventDetails, TrackEventSelection} from './selection';
-import {Ds} from '../trace_processor/dataset';
+import {Dataset} from '../trace_processor/dataset';
 
 export interface TrackManager {
   /**
@@ -179,7 +179,7 @@
    * Optional: Returns a dataset that represents the events displayed on this
    * track.
    */
-  getDataset?(): Ds.Dataset | undefined;
+  getDataset?(): Dataset | undefined;
 
   /**
    * Optional: Get details of a track event given by eventId on this track.
diff --git a/ui/src/trace_processor/dataset.ts b/ui/src/trace_processor/dataset.ts
index 6863570..25c64cb 100644
--- a/ui/src/trace_processor/dataset.ts
+++ b/ui/src/trace_processor/dataset.ts
@@ -16,71 +16,43 @@
 import {getOrCreate} from '../base/utils';
 import {ColumnType, SqlValue} from './query_result';
 
-export namespace Ds {
-  export type Dataset = UnionDataset | SourceDataset;
-  export type Schema = Record<string, ColumnType>;
-
+/**
+ * A dataset defines a set of rows in TraceProcessor and a schema of the
+ * resultant columns. Dataset implementations describe how to get the data in
+ * different ways - e.g. 'source' datasets define a dataset as a table name (or
+ * select statement) + filters, whereas a 'union' dataset defines a dataset as
+ * the union of other datasets.
+ *
+ * The idea is that users can build arbitrarily complex trees of datasets, then
+ * at any point call `optimize()` to create the smallest possible tree that
+ * represents the same dataset, and `query()` which produces a select statement
+ * for the resultant dataset.
+ *
+ * Users can also use the `schema` property and `implements()` to get and test
+ * the schema of a given dataset.
+ */
+export interface Dataset {
   /**
-   * Defines a dataset with a source SQL select statement of table name, a
-   * schema describing the columns, and an optional filter.
+   * Get or calculate the resultant schema of this dataset.
    */
-  export interface SourceDataset {
-    readonly src: string;
-    readonly schema: Schema;
-    readonly filter?: EqFilter | InFilter;
-  }
+  readonly schema: DatasetSchema;
 
   /**
-   * A dataset that represents the union of multiple datasets.
-   */
-  export interface UnionDataset {
-    readonly union: ReadonlyArray<Dataset>;
-  }
-
-  /**
-   * Generic filter type.
-   */
-  export type Filter = EqFilter | InFilter;
-
-  /**
-   * A filter used to express that a column must equal a value.
-   */
-  export interface EqFilter {
-    readonly col: string;
-    readonly eq: SqlValue;
-  }
-
-  /**
-   * A filter used to express that column must be one of a set of values.
-   */
-  export interface InFilter {
-    readonly col: string;
-    readonly in: ReadonlyArray<SqlValue>;
-  }
-
-  /**
-   * Returns true if the dataset implements a given schema.
+   * Produce a query for this dataset.
    *
-   * Note: `implements` is a reserved keyword in TS so we can't call this
-   * function `implements`.
-   *
-   * @param dataset - The dataset to test.
-   * @param testSchema - The schema to test against.
+   * @param schema - The schema to use for extracting columns - if undefined,
+   * the most specific possible schema is evaluated from the dataset first and
+   * used instead.
    */
-  export function doesImplement(dataset: Dataset, testSchema: Schema): boolean {
-    const datasetSchema = schema(dataset);
-    return Object.entries(testSchema).every(([name, kind]) => {
-      return name in datasetSchema && datasetSchema[name] === kind;
-    });
-  }
+  query(schema?: DatasetSchema): string;
 
   /**
-   * This function optimizes a dataset into the smallest possible expression.
+   * Optimizes a dataset into the smallest possible expression.
    *
    * For example by combining elements of union data sets that have the same src
    * and similar filters into a single set.
    *
-   * For example, the following union data set...
+   * For example, the following 'union' dataset...
    *
    * ```
    * {
@@ -105,7 +77,7 @@
    * }
    * ```
    *
-   * ...will be combined into a single set...
+   * ...will be combined into a single 'source' dataset...
    *
    * ```
    * {
@@ -117,142 +89,202 @@
    *   filter: {col: 'a', in: [1, 2]},
    * },
    * ```
-   *
-   * @param dataset - The dataset to optimize.
    */
-  export function optimize(dataset: Dataset): Dataset {
-    if ('src' in dataset) {
-      // No optimization possible for individual datasets
-      return dataset;
-    } else if ('union' in dataset) {
-      // Recursively optimize each dataset of this union
-      const optimizedUnion = dataset.union.map(optimize);
-
-      // Find all source datasets and combine then based on src
-      const combinedSrcSets = new Map<string, SourceDataset[]>();
-      const otherDatasets: Dataset[] = [];
-      for (const e of optimizedUnion) {
-        if ('src' in e) {
-          const set = getOrCreate(combinedSrcSets, e.src, () => []);
-          set.push(e);
-        } else {
-          otherDatasets.push(e);
-        }
-      }
-
-      const mergedSrcSets = Array.from(combinedSrcSets.values()).map(
-        (srcGroup) => {
-          if (srcGroup.length === 1) return srcGroup[0];
-
-          // Combine schema across all members in the union
-          const combinedSchema = srcGroup.reduce((acc, e) => {
-            Object.assign(acc, e.schema);
-            return acc;
-          }, {} as Schema);
-
-          // Merge filters for the same src
-          const inFilters: InFilter[] = [];
-          for (const {filter} of srcGroup) {
-            if (filter) {
-              if ('eq' in filter) {
-                inFilters.push({col: filter.col, in: [filter.eq]});
-              } else {
-                inFilters.push(filter);
-              }
-            }
-          }
-
-          const mergedFilter = mergeFilters(inFilters);
-          return {
-            src: srcGroup[0].src,
-            schema: combinedSchema,
-            filter: mergedFilter,
-          };
-        },
-      );
-
-      const finalUnion = [...mergedSrcSets, ...otherDatasets];
-
-      if (finalUnion.length === 1) {
-        return finalUnion[0];
-      } else {
-        return {union: finalUnion};
-      }
-    } else {
-      assertUnreachable(dataset);
-    }
-  }
-
-  function mergeFilters(filters: InFilter[]): InFilter | undefined {
-    if (filters.length === 0) return undefined;
-    const col = filters[0].col;
-    const values = new Set(filters.flatMap((filter) => filter.in));
-    return {col, in: Array.from(values)};
-  }
+  optimize(): Dataset;
 
   /**
-   * Get the schema of an dataset.
+   * Returns true if this dataset implements a given schema.
    *
-   * @param dataset - The dataset to get the schema of.
+   * @param schema - The schema to test against.
    */
-  export function schema(dataset: Dataset): Schema {
-    if ('src' in dataset) {
-      return dataset.schema;
-    } else if ('union' in dataset) {
-      // Find the minimal set of columns that are supported by all datasets of
-      // the union
-      let sch: Record<string, ColumnType> | undefined = undefined;
-      dataset.union.forEach((e) => {
-        const eSchema = schema(e);
-        if (sch === undefined) {
-          // First time just use this one
-          sch = eSchema;
-        } else {
-          const newSch: Record<string, ColumnType> = {};
-          for (const [key, kind] of Object.entries(sch)) {
-            if (key in eSchema && eSchema[key] === kind) {
-              newSch[key] = kind;
-            }
-          }
-          sch = newSch;
-        }
-      });
-      return sch ?? {};
-    } else {
-      assertUnreachable(dataset);
-    }
+  implements(schema: DatasetSchema): boolean;
+}
+
+/**
+ * Defines a list of columns and types that define the shape of the data
+ * represented by a dataset.
+ */
+export type DatasetSchema = Record<string, ColumnType>;
+
+/**
+ * A filter used to express that a column must equal a value.
+ */
+interface EqFilter {
+  readonly col: string;
+  readonly eq: SqlValue;
+}
+
+/**
+ * A filter used to express that column must be one of a set of values.
+ */
+interface InFilter {
+  readonly col: string;
+  readonly in: ReadonlyArray<SqlValue>;
+}
+
+/**
+ * Union of all filter types.
+ */
+type Filter = EqFilter | InFilter;
+
+/**
+ * Named arguments for a SourceDataset.
+ */
+interface SourceDatasetConfig {
+  readonly src: string;
+  readonly schema: DatasetSchema;
+  readonly filter?: Filter;
+}
+
+/**
+ * Defines a dataset with a source SQL select statement of table name, a
+ * schema describing the columns, and an optional filter.
+ */
+export class SourceDataset implements Dataset {
+  readonly src: string;
+  readonly schema: DatasetSchema;
+  readonly filter?: Filter;
+
+  constructor(config: SourceDatasetConfig) {
+    this.src = config.src;
+    this.schema = config.schema;
+    this.filter = config.filter;
   }
 
-  /**
-   * Produce a query for this dataset.
-   *
-   * @param dataset - The dataset to get the query for.
-   * @param sch - The schema to use for extracting columns - if undefined, the
-   * most specific possible schema is evaluated from the dataset first and used
-   * instead.
-   */
-  export function query(dataset: Dataset, sch?: Schema): string {
-    function filterToQuery(filter: Filter) {
-      if ('eq' in filter) {
-        return `where ${filter.col} = ${filter.eq}`;
-      } else if ('in' in filter) {
-        return `where ${filter.col} in (${filter.in.join(',')})`;
+  query(schema?: DatasetSchema) {
+    schema = schema ?? this.schema;
+    const cols = Object.keys(schema);
+    const whereClause = this.filterToQuery();
+    return `select ${cols.join(', ')} from (${this.src}) ${whereClause}`.trim();
+  }
+
+  optimize() {
+    // Cannot optimize SourceDataset
+    return this;
+  }
+
+  implements(schema: DatasetSchema) {
+    return Object.entries(schema).every(([name, kind]) => {
+      return name in this.schema && this.schema[name] === kind;
+    });
+  }
+
+  private filterToQuery() {
+    const filter = this.filter;
+    if (filter === undefined) {
+      return '';
+    }
+    if ('eq' in filter) {
+      return `where ${filter.col} = ${filter.eq}`;
+    } else if ('in' in filter) {
+      return `where ${filter.col} in (${filter.in.join(',')})`;
+    } else {
+      assertUnreachable(filter);
+    }
+  }
+}
+
+/**
+ * A dataset that represents the union of multiple datasets.
+ */
+export class UnionDataset implements Dataset {
+  constructor(readonly union: ReadonlyArray<Dataset>) {}
+
+  get schema(): DatasetSchema {
+    // Find the minimal set of columns that are supported by all datasets of
+    // the union
+    let sch: Record<string, ColumnType> | undefined = undefined;
+    this.union.forEach((ds) => {
+      const dsSchema = ds.schema;
+      if (sch === undefined) {
+        // First time just use this one
+        sch = dsSchema;
       } else {
-        assertUnreachable(filter);
+        const newSch: Record<string, ColumnType> = {};
+        for (const [key, kind] of Object.entries(sch)) {
+          if (key in dsSchema && dsSchema[key] === kind) {
+            newSch[key] = kind;
+          }
+        }
+        sch = newSch;
+      }
+    });
+    return sch ?? {};
+  }
+
+  query(schema?: DatasetSchema): string {
+    schema = schema ?? this.schema;
+    return this.union
+      .map((dataset) => dataset.query(schema))
+      .join(' union all ');
+  }
+
+  optimize(): Dataset {
+    // Recursively optimize each dataset of this union
+    const optimizedUnion = this.union.map((ds) => ds.optimize());
+
+    // Find all source datasets and combine then based on src
+    const combinedSrcSets = new Map<string, SourceDataset[]>();
+    const otherDatasets: Dataset[] = [];
+    for (const e of optimizedUnion) {
+      if (e instanceof SourceDataset) {
+        const set = getOrCreate(combinedSrcSets, e.src, () => []);
+        set.push(e);
+      } else {
+        otherDatasets.push(e);
       }
     }
 
-    sch = sch ?? schema(dataset);
-    if ('src' in dataset) {
-      const whereClause = dataset.filter ? filterToQuery(dataset.filter) : '';
-      const cols = Object.keys(sch);
-      return `select ${cols.join(', ')} from (${dataset.src}) ${whereClause}`.trim();
-    } else if ('union' in dataset) {
-      return dataset.union
-        .map((dataset) => query(dataset, sch))
-        .join(' union all ');
+    const mergedSrcSets = Array.from(combinedSrcSets.values()).map(
+      (srcGroup) => {
+        if (srcGroup.length === 1) return srcGroup[0];
+
+        // Combine schema across all members in the union
+        const combinedSchema = srcGroup.reduce((acc, e) => {
+          Object.assign(acc, e.schema);
+          return acc;
+        }, {} as DatasetSchema);
+
+        // Merge filters for the same src
+        const inFilters: InFilter[] = [];
+        for (const {filter} of srcGroup) {
+          if (filter) {
+            if ('eq' in filter) {
+              inFilters.push({col: filter.col, in: [filter.eq]});
+            } else {
+              inFilters.push(filter);
+            }
+          }
+        }
+
+        const mergedFilter = mergeFilters(inFilters);
+        return new SourceDataset({
+          src: srcGroup[0].src,
+          schema: combinedSchema,
+          filter: mergedFilter,
+        });
+      },
+    );
+
+    const finalUnion = [...mergedSrcSets, ...otherDatasets];
+
+    if (finalUnion.length === 1) {
+      return finalUnion[0];
     } else {
-      assertUnreachable(dataset);
+      return new UnionDataset(finalUnion);
     }
   }
+
+  implements(schema: DatasetSchema) {
+    return Object.entries(schema).every(([name, kind]) => {
+      return name in this.schema && this.schema[name] === kind;
+    });
+  }
+}
+
+function mergeFilters(filters: InFilter[]): InFilter | undefined {
+  if (filters.length === 0) return undefined;
+  const col = filters[0].col;
+  const values = new Set(filters.flatMap((filter) => filter.in));
+  return {col, in: Array.from(values)};
 }
diff --git a/ui/src/trace_processor/dataset_unittest.ts b/ui/src/trace_processor/dataset_unittest.ts
index e354b54..2bd4e53 100644
--- a/ui/src/trace_processor/dataset_unittest.ts
+++ b/ui/src/trace_processor/dataset_unittest.ts
@@ -12,168 +12,158 @@
 // See the License for the specific language governing permissions and
 // limitations under the License.
 
-import {Ds} from './dataset';
+import {SourceDataset, UnionDataset} from './dataset';
 import {LONG, NUM, STR} from './query_result';
 
 test('get query for simple dataset', () => {
-  const dataset: Ds.Dataset = {
+  const dataset = new SourceDataset({
     src: 'slice',
     schema: {id: NUM},
-  };
+  });
 
-  expect(Ds.query(dataset)).toEqual('select id from (slice)');
+  expect(dataset.query()).toEqual('select id from (slice)');
 });
 
 test("get query for simple dataset with 'eq' filter", () => {
-  const dataset: Ds.Dataset = {
+  const dataset = new SourceDataset({
     src: 'slice',
     schema: {id: NUM},
     filter: {
       col: 'id',
       eq: 123,
     },
-  };
+  });
 
-  expect(Ds.query(dataset)).toEqual('select id from (slice) where id = 123');
+  expect(dataset.query()).toEqual('select id from (slice) where id = 123');
 });
 
 test("get query for simple dataset with an 'in' filter", () => {
-  const dataset: Ds.Dataset = {
+  const dataset = new SourceDataset({
     src: 'slice',
     schema: {id: NUM},
     filter: {
       col: 'id',
       in: [123, 456],
     },
-  };
+  });
 
-  expect(Ds.query(dataset)).toEqual(
+  expect(dataset.query()).toEqual(
     'select id from (slice) where id in (123,456)',
   );
 });
 
 test('get query for union dataset', () => {
-  const dataset: Ds.Dataset = {
-    union: [
-      {
-        src: 'slice',
-        schema: {id: NUM},
-        filter: {
-          col: 'id',
-          eq: 123,
-        },
+  const dataset = new UnionDataset([
+    new SourceDataset({
+      src: 'slice',
+      schema: {id: NUM},
+      filter: {
+        col: 'id',
+        eq: 123,
       },
-      {
-        src: 'slice',
-        schema: {id: NUM},
-        filter: {
-          col: 'id',
-          eq: 456,
-        },
+    }),
+    new SourceDataset({
+      src: 'slice',
+      schema: {id: NUM},
+      filter: {
+        col: 'id',
+        eq: 456,
       },
-    ],
-  };
+    }),
+  ]);
 
-  expect(Ds.query(dataset)).toEqual(
+  expect(dataset.query()).toEqual(
     'select id from (slice) where id = 123 union all select id from (slice) where id = 456',
   );
 });
 
 test('doesImplement', () => {
-  const dataset = {
+  const dataset = new SourceDataset({
     src: 'slice',
     schema: {id: NUM, ts: LONG},
-  };
+  });
 
-  expect(Ds.doesImplement(dataset, {id: NUM})).toBe(true);
-  expect(Ds.doesImplement(dataset, {id: NUM, ts: LONG})).toBe(true);
-  expect(Ds.doesImplement(dataset, {id: NUM, ts: LONG, name: STR})).toBe(false);
-  expect(Ds.doesImplement(dataset, {id: LONG})).toBe(false);
+  expect(dataset.implements({id: NUM})).toBe(true);
+  expect(dataset.implements({id: NUM, ts: LONG})).toBe(true);
+  expect(dataset.implements({id: NUM, ts: LONG, name: STR})).toBe(false);
+  expect(dataset.implements({id: LONG})).toBe(false);
 });
 
 test('find the schema of a simple dataset', () => {
-  const dataset: Ds.Dataset = {
+  const dataset = new SourceDataset({
     src: 'slice',
     schema: {id: NUM, ts: LONG},
-  };
+  });
 
-  expect(Ds.schema(dataset)).toMatchObject({id: NUM, ts: LONG});
+  expect(dataset.schema).toMatchObject({id: NUM, ts: LONG});
 });
 
 test('find the schema of a union where source sets differ in their names', () => {
-  const dataset: Ds.Dataset = {
-    union: [
-      {
-        src: 'slice',
-        schema: {foo: NUM},
-      },
-      {
-        src: 'slice',
-        schema: {bar: NUM},
-      },
-    ],
-  };
+  const dataset = new UnionDataset([
+    new SourceDataset({
+      src: 'slice',
+      schema: {foo: NUM},
+    }),
+    new SourceDataset({
+      src: 'slice',
+      schema: {bar: NUM},
+    }),
+  ]);
 
-  expect(Ds.schema(dataset)).toMatchObject({});
+  expect(dataset.schema).toMatchObject({});
 });
 
 test('find the schema of a union with differing source sets', () => {
-  const dataset: Ds.Dataset = {
-    union: [
-      {
-        src: 'slice',
-        schema: {foo: NUM},
-      },
-      {
-        src: 'slice',
-        schema: {foo: LONG},
-      },
-    ],
-  };
+  const dataset = new UnionDataset([
+    new SourceDataset({
+      src: 'slice',
+      schema: {foo: NUM},
+    }),
+    new SourceDataset({
+      src: 'slice',
+      schema: {foo: LONG},
+    }),
+  ]);
 
-  expect(Ds.schema(dataset)).toMatchObject({});
+  expect(dataset.schema).toMatchObject({});
 });
 
 test('find the schema of a union with one column in common', () => {
-  const dataset: Ds.Dataset = {
-    union: [
-      {
-        src: 'slice',
-        schema: {foo: NUM, bar: NUM},
-      },
-      {
-        src: 'slice',
-        schema: {foo: NUM, baz: NUM},
-      },
-    ],
-  };
+  const dataset = new UnionDataset([
+    new SourceDataset({
+      src: 'slice',
+      schema: {foo: NUM, bar: NUM},
+    }),
+    new SourceDataset({
+      src: 'slice',
+      schema: {foo: NUM, baz: NUM},
+    }),
+  ]);
 
-  expect(Ds.schema(dataset)).toMatchObject({foo: NUM});
+  expect(dataset.schema).toMatchObject({foo: NUM});
 });
 
 test('optimize a union dataset', () => {
-  const dataset: Ds.Dataset = {
-    union: [
-      {
-        src: 'slice',
-        schema: {},
-        filter: {
-          col: 'track_id',
-          eq: 123,
-        },
+  const dataset = new UnionDataset([
+    new SourceDataset({
+      src: 'slice',
+      schema: {},
+      filter: {
+        col: 'track_id',
+        eq: 123,
       },
-      {
-        src: 'slice',
-        schema: {},
-        filter: {
-          col: 'track_id',
-          eq: 456,
-        },
+    }),
+    new SourceDataset({
+      src: 'slice',
+      schema: {},
+      filter: {
+        col: 'track_id',
+        eq: 456,
       },
-    ],
-  };
+    }),
+  ]);
 
-  expect(Ds.optimize(dataset)).toEqual({
+  expect(dataset.optimize()).toEqual({
     src: 'slice',
     schema: {},
     filter: {
@@ -184,28 +174,26 @@
 });
 
 test('optimize a union dataset with different types of filters', () => {
-  const dataset: Ds.Dataset = {
-    union: [
-      {
-        src: 'slice',
-        schema: {},
-        filter: {
-          col: 'track_id',
-          eq: 123,
-        },
+  const dataset = new UnionDataset([
+    new SourceDataset({
+      src: 'slice',
+      schema: {},
+      filter: {
+        col: 'track_id',
+        eq: 123,
       },
-      {
-        src: 'slice',
-        schema: {},
-        filter: {
-          col: 'track_id',
-          in: [456, 789],
-        },
+    }),
+    new SourceDataset({
+      src: 'slice',
+      schema: {},
+      filter: {
+        col: 'track_id',
+        in: [456, 789],
       },
-    ],
-  };
+    }),
+  ]);
 
-  expect(Ds.optimize(dataset)).toEqual({
+  expect(dataset.optimize()).toEqual({
     src: 'slice',
     schema: {},
     filter: {
@@ -216,20 +204,18 @@
 });
 
 test('optimize a union dataset with different schemas', () => {
-  const dataset: Ds.Dataset = {
-    union: [
-      {
-        src: 'slice',
-        schema: {foo: NUM},
-      },
-      {
-        src: 'slice',
-        schema: {bar: NUM},
-      },
-    ],
-  };
+  const dataset = new UnionDataset([
+    new SourceDataset({
+      src: 'slice',
+      schema: {foo: NUM},
+    }),
+    new SourceDataset({
+      src: 'slice',
+      schema: {bar: NUM},
+    }),
+  ]);
 
-  expect(Ds.optimize(dataset)).toEqual({
+  expect(dataset.optimize()).toEqual({
     src: 'slice',
     // The resultant schema is the combination of the union's member's schemas,
     // as we know the source is the same as we know we can get all of the 'seen'