Broker.get_table(headers, stream_name='primary', fields=None, fill=False, handler_registry=None, convert_times=True, timezone=None, localize_times=True)[source]

Load the data from one or more runs as a table (pandas.DataFrame).

headersHeader or iterable of Headers

The headers to fetch the events for

stream_namestr, optional

Get events from only “event stream” with this name.

Default is ‘primary’

fieldsList[str], optional

whitelist of field names of interest; if None, all are returned

Default is None

fillbool or Iterable[str], optional

Which fields to fill. If True, fill all possible fields.

Each event will have the data filled for the intersection of it’s external keys and the fields requested filled.

Default is False

handler_registrydict, optional

mapping filestore specs (strings) to handlers (callable classes)

convert_timesbool, optional

Whether to convert times from float (seconds since 1970) to numpy datetime64, using pandas. True by default.

timezonestr, optional

e.g., ‘US/Eastern’; if None, use metadatastore configuration in self.mds.config[‘timezone’]

handler_registrydict, optional

mapping asset specs (strings) to handlers (callable classes)

localize_timesbool, optional

If the times should be localized to the ‘local’ time zone. If True (the default) the time stamps are converted to the localtime zone (as configure in mds).

This is problematic for several reasons:

  • apparent gaps or duplicate times around DST transitions

  • incompatibility with every other time stamp (which is in UTC)

however, this makes the dataframe repr look nicer

This implies convert_times.

Defaults to True to preserve back-compatibility.